UTDrive: Driver Behavior and Speech Interactive Systems for In-Vehicle Environments

This paper describes an overview of the UTDrive project. UTDrive is part of an on-going international collaboration to collect and research rich multi-modal data recorded for modeling driver behavior for in-vehicle environments. The objective of the UTDrive project is to analyze behavior while the driver is interacting with speech-activated systems or performing common secondary tasks, as well as to better understand speech characteristics of the driver undergoing additional cognitive load. The corpus consists of audio, video, gas/brake pedal pressure, forward distance, GPS information, and CAN-Bus information. The resulting corpus, analysis, and modeling will contribute to more effective speech interactive systems with are less distractive and adjustable to the driver's cognitive capacity and driving situations.

[1]  H L HansenJohn Analysis and compensation of speech under stress and noise for environmental robustness in speech recognition , 1996 .

[2]  John H. L. Hansen,et al.  Audio-visual SPeaker localization for car navigation systems , 2004, INTERSPEECH.

[3]  Alex Pentland,et al.  Graphical models for driver behavior recognition in a SmartCar , 2000, Proceedings of the IEEE Intelligent Vehicles Symposium 2000 (Cat. No.00TH8511).

[4]  Bhiksha Raj,et al.  A Comparison Between Spoken Queries and Menu-Based Interfaces for In-car Digital Music Selection , 2005, INTERACT.

[5]  Robert Graham,et al.  Experimental Comparison of Manual and Voice Controls for the Operation of in-Vehicle Systems , 2000 .

[6]  John H. L. Hansen,et al.  CSA-BF: a constrained switched adaptive beamformer for speech enhancement and recognition in real car environments , 2003, IEEE Trans. Speech Audio Process..

[7]  Hong-Seok Kim,et al.  Performance of an HMM speech recognizer using a real-time tracking microphone array as input , 1999, IEEE Trans. Speech Audio Process..

[8]  Mikko Sams Audio-visual speech processing and attention , 2003 .

[9]  Hakan Erdogan,et al.  Multi-modal Person Recognition for Vehicular Applications , 2005, Multiple Classifier Systems.

[10]  Tsuhan Chen,et al.  Audiovisual speech processing , 2001, IEEE Signal Process. Mag..

[11]  John H. L. Hansen,et al.  "CU-move": robust speech processing for in-vehicle speech systems , 2000, INTERSPEECH.

[12]  Ming Liu,et al.  AVICAR: audio-visual speech corpus in a car environment , 2004, INTERSPEECH.

[13]  Paul Green,et al.  Safety and Usability of Speech Interfaces for In-Vehicle Tasks while Driving: A Brief Literature Review , 2006 .

[14]  Kazuya Takeda,et al.  Driver Recognition System Using FNN and Statistical Methods , 2007 .

[15]  John H. L. Hansen,et al.  Analysis and compensation of speech under stress and noise for environmental robustness in speech recognition , 1996, Speech Commun..

[16]  Alex Pentland,et al.  Modeling and Prediction of Human Behavior , 1999, Neural Computation.

[17]  Rosalind W. Picard,et al.  Modeling drivers' speech under stress , 2003, Speech Commun..