UTDrive: The Smart Vehicle Project

This chapter presents research activities of UTDrive: the smart vehicle project, at the Center for Robust Speech Systems, University of Texas at Dallas. The objectives of the UTDrive project are to collect and research rich multi-modal data recorded in actual car environments for analyzing and modeling driver behavior. The models of driver behavior under normal and distracted driving conditions can be used to create improved in-vehicle human–machine interactive systems and reduce vehicle accidents on the road. The UTDrive corpus consists of audio, video, brake/gas pedal pressure, head distance, GPS information (e.g., position, velocity), and CAN-bus information (e.g., steering-wheel angle, brake position, throttle position, and vehicle speed). Here, we describe our in-vehicle data collection framework, data collection protocol, dialog and secondary task demands, data analysis, and preliminary experimental results. Finally, we discuss our proposed multi-layer data transcription procedure for in-vehicle data collection and future research directions.

[1]  Kazuya Takeda,et al.  On-Going Data Collection of Driving Behavior Signals , 2009 .

[2]  Tsuhan Chen,et al.  Audiovisual speech processing , 2001, IEEE Signal Process. Mag..

[3]  Erwin R. Boer,et al.  Toward an Integrated Model of Driver Behavior in Cognitive Architecture , 2001 .

[4]  Paul Green,et al.  Safety and Usability of Speech Interfaces for In-Vehicle Tasks while Driving: A Brief Literature Review , 2006 .

[5]  John H. L. Hansen,et al.  Audio-visual SPeaker localization for car navigation systems , 2004, INTERSPEECH.

[6]  Bhiksha Raj,et al.  A Comparison Between Spoken Queries and Menu-Based Interfaces for In-car Digital Music Selection , 2005, INTERACT.

[7]  Hakan Erdogan,et al.  Multi-modal Person Recognition for Vehicular Applications , 2005, Multiple Classifier Systems.

[8]  Hong-Seok Kim,et al.  Performance of an HMM speech recognizer using a real-time tracking microphone array as input , 1999, IEEE Trans. Speech Audio Process..

[9]  Gary Burnett,et al.  Defining Driver Distraction , 2005 .

[10]  Volkan Sezer,et al.  Data collection with "UYANIK": too much pain; but gains are coming , 2007 .

[11]  Kazuya Takeda,et al.  Driver Modeling Based on Driving Behavior and Its Evaluation in Driver Identification , 2007, Proceedings of the IEEE.

[12]  Kazuya Takeda,et al.  Driver Recognition System Using FNN and Statistical Methods , 2007 .

[13]  John H. L. Hansen,et al.  Analysis and compensation of speech under stress and noise for environmental robustness in speech recognition , 1996, Speech Commun..

[14]  John H. L. Hansen,et al.  ICARUS: Source generator based real-time recognition of speech in noisy stressful and Lombard effect environments , 1995, Speech Commun..

[15]  Robert Graham,et al.  Experimental Comparison of Manual and Voice Controls for the Operation of in-Vehicle Systems , 2000 .

[16]  John H. L. Hansen,et al.  CSA-BF: a constrained switched adaptive beamformer for speech enhancement and recognition in real car environments , 2003, IEEE Trans. Speech Audio Process..

[17]  Alex Pentland,et al.  Modeling and Prediction of Human Behavior , 1999, Neural Computation.