Research Paper On Artificial Passenger

 Download your Reports for Artificial Passenger

The AP is an artificial intelligence?based companion that will be resident in software and chips embedded in the automobile dashboard. The heart of the system is a conversation planner that holds a profile of you, including details of your interests and profession.

????????????? A microphone picks up your answer and breaks it down into separate words with speech-recognition software. A camera built into the dashboard also tracks your lip movements to improve the This research suggests that we can make predictions about various aspects of driver performance based on what we glean from the movements of a driver?s eyes and that a system can eventually be developed to capture this data and use it to alert people when their driving has become significantly impaired by fatigue.

???????????????????????????? The natural dialog car system analyzes a driver?s answer and the contents of the answer together with his voice patterns to determine if he is alert while driving. The system warns the driver or changes the topic of conversation if the system determines that the driver is about to fall asleep. The system may also detect whether a
driver is affected by alcohol or drugs.

In this seminar , we are giving some basic concepts about smart cards. An artificial passenger (AP) is a device that would be used in a motor vehicle to make sure that the driver stays awake. IBM has developed a prototype that holds a conversation with a driver, telling jokes and asking questions intended to determine whether the driver can respond alertly enough. Assuming the IBM approach, an artificial passenger would use a microphone for the driver and a speech generator and the vehicle?s audio speakers to converse with the driver. The conversation would be based on a personalized profile of the driver. A camera could be used to evaluate the driver?s ?facial state? and a voice analyzer to evaluate whether the driver was becoming drowsy. If a driver seemed to display too much fatigue, the artificial passenger might be programmed to open all the windows, sound a buzzer, increase background music volume, or even spray the driver with ice water. One of the ways to address driver safety concerns is to develop an efficient system that relies on voice instead of hands to control Telematics devices. It has been shown in various experiments that well designed voice control interfaces can reduce a driver?s distraction compared with manual control situations. One of the ways to reduce a driver?s cognitive workload is to allow the driver to speak naturally when interacting with a car system (e.g.when playing voice games, issuing commands via voice). It is difficult for a driver to remember a syntax, such as ?What is the distance to JFK??"Or how far is JFK?? or ?How long to drive to JFK?? etc.). This fact led to the development of Conversational Interactivity for Telematics (CIT) speech systems at IBM Research.
CIT speech systems can significantly improve a driver-vehicle relationship and contribute to driving safety. But the development of full fledged Natural Language Understanding (NLU) for CIT is a difficult problem that typically requires significant computer resources that are usually not available in local computer processors that car manufacturer provide for their cars.
To address this, NLU components should be located on a server that is accessed by cars remotely or NLU should be downsized to run on local computer devices (that are typically based on embedded chips).Some car manufacturers see advantages in using upgraded NLU and speech processing on the client in the car, since remote connections to servers are not available everywhere, can have delays, and are no trobust. Our department is developing a ?quasi-NLU?component ? a ?reduced? variant of NLU that can be run in CPU systems with relatively limited resources.

 Download your Reports for Artificial Passenger


Sponsored Links:-

  • Amir, N., & Ron, S. (1998) Towards an automatic classification of emotions in speech. Proceedings of the 5th International Conference of Spoken Language Processing (pp. 555–558).Google Scholar

  • Bahl, L. R., et al. (1995). Performance of the IBM large vocabulary continuous speech recognition system on the ARPA Wall Street Journal Task. International Conference on Acoustics, Speech and Signal Processing, 1, 41-44.Google Scholar

  • Bahl, L. R., Brown, P. F., de Souza, P. V., & Mercer, R. L. (1989). A tree-based statistical language model for natural language speech recognition. IEEE Transactions on Acoustics, Speech and Signal Processing, 37(7), 1001-1008.CrossRefGoogle Scholar

  • Bahl, L. R., de Souza, P. V., Gopalakrishnan, P. S., Nahamoo, D., & Picheny, M. (1994). Robust methods for using context-dependent features and models in continuous speech recognizer. Proceedings of the International Conference on Acoustics, Speech and Signal Processin, I, 533-536.Google Scholar

  • Bahl, L. R., de Souza, P. V., Gopalakrishnan, P. S., Nahamoo, D., & Picheny, M. (1991). Decision trees for phonological rules in continuous speech. Proceedings of the 1991 International Conference on Acoustic, Speech, and Signal Processing(pp.185-188). Toronto, Canada.Google Scholar

  • Bain, K., Basson, S., Faisman, A., & Kanevsky, D. (2005). Accessibility, transcription, and access everywhere, IBM Systems Journal, 44(3), 589-604.CrossRefGoogle Scholar

  • Basson, S., Fairweather, P., & Kanevsky, D. (2003). Artificial passenger with condition sensors. U.S. Patent No. 6,792,339.Google Scholar

  • Basson, S. Kanevsky, D., Sabath, M., Sedivy, J., & Zlatsin, A. (2001). Virtual cooperative network formed by local clients in zones without cellular services. U.S. Patent No. 6,965,773.Google Scholar

  • Bellegarda, J. R. (1996). Context-dependent vector clustering for speech recognition. In C-H. Lee and F. K. Song (Eds.), Automatic speech and speaker recognition (pp. 133-153), Boston: Kluwer Academic Publishers.Google Scholar

  • Comerford, L., Frank, D., Gopalakrsihnan, P., Gopinath, R., & Sedivy, J. (2001). The IBM personal Speech Assistant. Proceedings of the International Conference on Acoustics, Speech and Signal Processing.Google Scholar

  • Deligne, S., Eide, E., Gopinath, R., Kanevsky, D., Maison, B., Olsen, P., Printz, H., & Sedivy, J. (2001). Low-resource speech recognition of 500-word vocabularies. Proceedings of EUROSPEECH 01.Google Scholar

  • Eisenberg, A. (2001, December 27). A passenger whose chatter is always appreciated, The New York Times. Scholar

  • Ferguson, G., Allen J. F., Miller, B. W. & Ringger, E. K. (1996). The design and implementation of the TRAINS-96 system: A prototype mixed-initiative planning assistant (TRAINS Technical Note 96-5). Rochester, NY: The University of Rochester, Computer Science Department.Google Scholar

  • Gales, M. J. F., & Woodland, P.C. (1996). Mean and variance adaptation within the MLLR framework. Computer Speech and Language, 10(4), 249-264.CrossRefGoogle Scholar

  • Gazdar G., & Mellish, C. (1989). Natural language processing in POP-11. New York: Addison-Wesley.Google Scholar

  • Gellatly, A. W., & Dingus, T. A. (1998). Speech recognition and automotive applications: Using speech to perform in-vehicle tasks. Proceedings of the Human Factors and Ergonomics Society 42nd Annual Meeting(pp. 1247-1251). Santa Monica, CA: Human Factors and Ergonomics Society.Google Scholar

  • Global System for Mobile Comminications (2004). Recognition performance evaluations of codecs for Speech Enabled Services” (SES) (Release 6) (Technical Report 3GPP TR 26.943 v6.00 (2004-12)). Scholar

  • Gopalakrishnan, P. S., Kanevsky, D., N’adas, A., & Nahamoo, D. (1991). An inequality for rational functions with applications to some statistical estimation problems IEEE Transactions on Information Theory, 37(1), 107-113.MATHCrossRefGoogle Scholar

  • Gould, L. S. (2003). IBM’s near-future technologies for telematics. In Automotive design and production. Scholar

  • Green, M., & Senders, J. (2004) Human error in road accidents. Scholar

  • IBM – Haifa. (2007). .Google Scholar

  • Jelinek, F. (1999). Statistical methods for speech recognition. Cambridge, MA: MIT Press.Google Scholar

  • Jelinek, F., et al. (1994). Decision tree parsing using a hidden derivation model. Proceedings of the ARPA Workshop on Human Language Technology, Princeton, N.J.Google Scholar

  • Kanevsky, D. (2007). Predicitive user modeling (IBM Research Report).Google Scholar

  • Kanevsky, D. (2007). Evaluating/predicting consequences of wrong recognition(IBM Research Report).Google Scholar

  • Kanevsky, D., Churchill, B., Faisman, A., Sicconi, R., & Nahamoo, D. (2004). Driver safety manager Proceedings of SPECOM 2004. .Google Scholar

  • Kanevsky, D., Libal, V., Sedivy, J., Zadrozny, W. (2002). Speaker model adaptation via network of similar users. U.S. Patent: No 6,442,519.Google Scholar

  • Kanevsky, D., & Maes, S. (2002). Apparatus and methods for user recognition employing behavioral passwords, U.S. Patent No. 6,421,453.Google Scholar

  • Kanevsky, D. R., Sicconi, R., & Viswanathan, M. (2004). Touch gesture based interface for motor vehicle. U.S. Patent application 20060047386, filed August 31, 2004.Google Scholar

  • Kanevsky, D., Maes, S., Sorensen, H., & Scott, J. (2003). Conversational data mining. U.S. Patent No. 6,665,644.Google Scholar

  • Kanevsky, D., Samdani, T., Zadrozny, W., & Zlatsin, A. (2003). System and method for resolving decoding ambiguity via dialog. U.S. Patent No. 6,587,818.Google Scholar

  • Kanevsky, D., Wolf, C., & Zadrozny, W. (2002). Language model adaptation via network of similar users. U.S. Patent: No. 6,484,136.Google Scholar

  • Kanevsky, D., & Yashchin, E. (2003). Method and system for off-line detection of textual topical changes and topic identification via likelihood based methods for improved language modeling. U.S. Patent: No. 6,529,902.Google Scholar

  • Kanevsky, D. & Zadrozny, W. (2001). Sleep prevention dialog based car system. U.S. Patent: No. 6,236,968.Google Scholar

  • Kanevsky, D. Zlatsin, A. (2001). Distributed personalized advertisement system and method. U.S. Patent: No 6,334,109.Google Scholar

  • Kanevsky, D., Zlatsin, A., & Pickover, C. (2003). Educational monitoring method and system for improving interactive skills based on participants on the network. U.S. Patent: No. 6,505,208.Google Scholar

  • Kensinger E. A., & Corkin, S. (2004). Two routes to emotional memory: Distinct neural processes for valence and arousal, PNAS, 101(9), 3310-3315. Scholar

  • Lee, C.H., & Gauvain, J. L. (1996). Bayesian adaptive learning and MAP estimation of HMM. In C-H. Lee and F. K. Song (Eds.), Automatic speech and speaker recognition(pp. 109-132). Boston: Kluwer Academic Publishers.Google Scholar

  • Lehnert, W., Dyer, M., Johnson, P., Yang, C. J., & Harley, S. (1983). BORIS–An experiment in-depth understanding of narratives, Artificial Intelligence, 20(1), 15-62.CrossRefGoogle Scholar

  • Lincoln, M., Cox, S., & Ringland, S. (1998). A comparison of two unsupervised approaches to accent identification. Proceedings of the 1998 International Conference on Spoken Language Processing (ICSLP 98)(pp. 317-320). Sydney, Australia.Google Scholar

  • Mangu, L., Brill, E., & Stolcke, A. (1999) Finding consensus among words: Lattice-based word error minimization. Proceedings of EUROSPEECH99.Google Scholar

  • Matrouf, D., Adda-Decker, M., Lamel L., & Gauvain, J. (1998). Language identification incorporating lexical information. Proceedings of the 1998 International Conference on Spoken Language Processing (ICSLP 98)(pp.181-184) Sydney, Australia.Google Scholar

  • McCallum, M. C., Campbell, J. L., Richman, J. B., & Brown, J. L. (2004). Speech recognition and in-vehicle telematics devices: Potential reductions in driver distraction. International Journal of Speech Technology, 7(1), 25-33.CrossRefGoogle Scholar

  • National Highway Traffic Safety Administration (NHTSA). (2007). gov/departments/nrd-13/DriverDistraction.html.Google Scholar

  • Normandin, Y. (1991). Hidden Markov models, maximum mutual information estimation and the speech recognition problem. Unpublished Ph.D. thesis, McGill University, Montreal, Canada.Google Scholar

  • Pereira, C., & Watson, C. (1998). Some acoustic characteristics of emotion. Proceedings of the 5th International Conference of Spoken Language Processing (pp. 927-930).Google Scholar

  • Picard R. W., & Scheirer, J. (2001). The Galvactivator: A glove that senses and communicates skin conductivity, MIT Media Laboratory. tech-reports/TR-542.pdf .Google Scholar

  • Rosetta Consortium. (2007). references.htm.Google Scholar

  • Sample, I. (2001). You drive me crazy. New Scientist. mg17123002.900.html .Google Scholar

  • Sicconi, R., & Lai, J. (2007). Speech interface in telematics,Unpublished paper.Google Scholar

  • Siebert, K-W. (2003). European telecommunications market trends: Future challenges for mobile service providers.{%}20Siebert.pdf .Google Scholar

  • Treat, J. R., Tumbas, N. S., McDonald, S. T., Shinar, D., Hume, R. D., Mayer, R. E., Stanisfer, R.L. & Castellan, N. J. (1977), Tri-level study of the causes of traffic accidents (U. S. Department of Transportation Report No. DOT-HS-034-3-535-77 (TAC).Google Scholar

  • Ward, M. (2001). Smart dashboard watches drivers, BBC News. science/nature/1445342.stm .Google Scholar

  • Whiteside, S. P. (1998). Simulated emotions: An acoustic study of voice and perturbation measures. Proceedings of the 5th International Conference of Spoken Language Processing (pp. 699-703).Google Scholar

  • One thought on “Research Paper On Artificial Passenger

    Leave a Reply

    Your email address will not be published. Required fields are marked *