Driver activity monitoring through supervised and unsupervised learning

This paper presents two different learning methods applied to the task of driver activity monitoring. The goal of the methods is to detect periods of driver activity that are not safe, such as talking on a cellular telephone, eating, or adjusting the dashboard radio system. The system presented here uses a side-mounted camera looking at a driver's profile and utilizes the silhouette appearance obtained from skin-color segmentation for detecting the activities. The unsupervised method uses agglomerative clustering to succinctly represent driver activities throughout a sequence, while the supervised learning method uses a Bayesian eigen-image classifier to distinguish between activities. The results of the two learning methods applied to driving sequences on three different subjects are presented and extensively discussed.

[1]  Mubarak Shah,et al.  Determining driver visual attention with one camera , 2003, IEEE Trans. Intell. Transp. Syst..

[2]  Wang Rongben,et al.  Driver's eye state detecting method design based on eye geometry feature , 2004, IEEE Intelligent Vehicles Symposium, 2004.

[3]  Youding Zhu,et al.  Head pose estimation for driver monitoring , 2004, IEEE Intelligent Vehicles Symposium, 2004.

[4]  Robert T. Collins,et al.  Articulated Motion Modeling for Activity Analysis , 2004, 2004 Conference on Computer Vision and Pattern Recognition Workshop.

[5]  Osama Masoud,et al.  Vision-based methods for driver monitoring , 2003, Proceedings of the 2003 IEEE International Conference on Intelligent Transportation Systems.

[6]  Alex Pentland,et al.  Pfinder: real-time tracking of the human body , 1996, Proceedings of the Second International Conference on Automatic Face and Gesture Recognition.

[7]  Simon Baker Real-time non-rigid driver head tracking for driver mental state estimation , 2004 .

[8]  Yang Song,et al.  Learning probabilistic structure for human motion detection , 2001, Proceedings of the 2001 IEEE Computer Society Conference on Computer Vision and Pattern Recognition. CVPR 2001.

[9]  David J. Kriegman,et al.  Recognition using class specific linear projection , 1997 .

[10]  Mikhail Belkin,et al.  Laplacian Eigenmaps for Dimensionality Reduction and Data Representation , 2003, Neural Computation.

[11]  Zhiwei Zhu,et al.  Real-time eye detection and tracking under various light conditions , 2002, ETRA.

[12]  James W. Davis,et al.  The Recognition of Human Movement Using Temporal Templates , 2001, IEEE Trans. Pattern Anal. Mach. Intell..

[13]  S T Roweis,et al.  Nonlinear dimensionality reduction by locally linear embedding. , 2000, Science.

[14]  Larry S. Davis,et al.  W4: Real-Time Surveillance of People and Their Activities , 2000, IEEE Trans. Pattern Anal. Mach. Intell..

[15]  Alex Pentland,et al.  Bayesian face recognition , 2000, Pattern Recognit..

[16]  Shumeet Baluja,et al.  Non-Intrusive Gaze Tracking Using Artificial Neural Networks , 1993, NIPS.

[17]  Marian Stewart Bartlett,et al.  Face recognition by independent component analysis , 2002, IEEE Trans. Neural Networks.

[18]  M. Turk,et al.  Eigenfaces for Recognition , 1991, Journal of Cognitive Neuroscience.

[19]  Vladimir Vezhnevets,et al.  A Survey on Pixel-Based Skin Color Detection Techniques , 2003 .

[20]  David J. Kriegman,et al.  Eigenfaces vs. Fisherfaces: Recognition Using Class Specific Linear Projection , 1996, ECCV.

[21]  Fengliang Xu,et al.  Real-time eye detection and tracking for driver observation under various light conditions , 2002, Intelligent Vehicle Symposium, 2002. IEEE.

[22]  Qiang Ji,et al.  Real-Time Eye, Gaze, and Face Pose Tracking for Monitoring Driver Vigilance , 2002, Real Time Imaging.