Gaze-based Intention Anticipation over Driving Manoeuvres in Semi-Autonomous Vehicles

Anticipating a human collaborator’s intention enables safe and efficient interaction between a human and an autonomous system. Specifically, in the context of semiautonomous driving, studies have revealed that correct and timely prediction of the driver’s intention needs to be an essential part of Advanced Driver Assistance System (ADAS) design. To this end, we propose a framework that exploits drivers’ time-series eye gaze and fixation patterns to anticipate their real-time intention over possible future manoeuvres, enabling a smart and collaborative ADAS that can aid drivers to overcome safety-critical situations. The method models human intention as the latent states of a hidden Markov model and uses probabilistic dynamic time warping distributions to capture the temporal characteristics of the observation patterns of the drivers. The method is evaluated on a data set of 124 experiments from 75 drivers collected in a safety-critical semi-autonomous driving scenario. The results illustrate the efficacy of the framework by correctly anticipating the drivers’ intentions about 3 seconds beforehand with over 90% accuracy.

[1]  Lina Yao,et al.  TagFall: Towards Unobstructive Fine-Grained Fall Detection based on UHF Passive RFID Tags , 2015, MobiQuitous.

[2]  Tarak Gandhi,et al.  Looking-In and Looking-Out of a Vehicle: Computer-Vision-Based Enhanced Vehicle Safety , 2007, IEEE Transactions on Intelligent Transportation Systems.

[3]  Lei Yang,et al.  Device-free human localization and tracking with UHF passive RFID tags: A data-driven approach , 2018, J. Netw. Comput. Appl..

[4]  Yiannis Demiris,et al.  Distributed, predictive perception of actions: a biologically inspired robotics architecture for imitation and learning , 2003, Connect. Sci..

[5]  Joseph B. Kruskall,et al.  The Symmetric Time-Warping Problem : From Continuous to Discrete , 1983 .

[6]  Bilge Mutlu,et al.  Anticipatory robot control for efficient human-robot collaboration , 2016, 2016 11th ACM/IEEE International Conference on Human-Robot Interaction (HRI).

[7]  Sung Kyung Hong,et al.  Real-time categorization of driver's gaze zone using the deep learning techniques , 2016, 2016 International Conference on Big Data and Smart Computing (BigComp).

[8]  Dana H. Ballard,et al.  Recognizing Behavior in Hand-Eye Coordination Patterns , 2009, Int. J. Humanoid Robotics.

[9]  Hema Swetha Koppula,et al.  Car that Knows Before You Do: Anticipating Maneuvers via Learning Temporal Driving Models , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).

[10]  Natasha Merat,et al.  Are you in the loop? Using gaze dispersion to understand driver visual attention during vehicle automation , 2017 .

[11]  G. Csibra,et al.  'Obsessed with goals': functions and mechanisms of teleological interpretation of actions in humans. , 2007, Acta psychologica.

[12]  Lina Yao,et al.  Device-free indoor localization and tracking through Human-Object Interactions , 2016, 2016 IEEE 17th International Symposium on A World of Wireless, Mobile and Multimedia Networks (WoWMoM).

[13]  Karen M. Feigh,et al.  Learning to Predict Intent from Gaze During Robotic Hand-Eye Coordination , 2017, AAAI.

[14]  Chrystopher L. Nehaniv,et al.  Imitation as a Dual-Route Process Featuring Predictive and Learning Components: A Biologically Plausible Computational Model , 2002 .

[15]  Mohan M. Trivedi,et al.  On-road prediction of driver's intent with multimodal sensory cues , 2011, IEEE Pervasive Computing.

[16]  Natasha Merat,et al.  Were they in the loop during automated driving? Links between visual attention and crash potential , 2016, Injury Prevention.

[17]  G. Aschersleben,et al.  The Theory of Event Coding (TEC): a framework for perception and action planning. , 2001, The Behavioral and brain sciences.

[18]  Dario D. Salvucci Inferring Driver Intent: A Case Study in Lane-Change Detection , 2004 .

[19]  Tanja Schultz,et al.  Combined intention, activity, and motion recognition for a humanoid household robot , 2011, 2011 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[20]  Meinard Müller,et al.  Information retrieval for music and motion , 2007 .

[21]  S. Baron-Cohen,et al.  The "Reading the Mind in the Eyes" Test revised version: a study with normal adults, and adults with Asperger syndrome or high-functioning autism. , 2001, Journal of child psychology and psychiatry, and allied disciplines.

[22]  Ashwin P. Dani,et al.  Bayesian human intention inference through multiple model filtering with gaze-based priors , 2016, 2016 19th International Conference on Information Fusion (FUSION).

[23]  Bilge Mutlu,et al.  Using gaze patterns to predict task intent in collaboration , 2015, Front. Psychol..

[24]  Simo Srkk,et al.  Bayesian Filtering and Smoothing , 2013 .

[25]  Natasha Merat,et al.  Coming back into the loop: Drivers' perceptual-motor performance in critical events after automated driving. , 2017, Accident; analysis and prevention.

[26]  B. Malle,et al.  The Folk Concept of Intentionality , 1997 .

[27]  Mohan M. Trivedi,et al.  On generalizing driver gaze zone estimation using convolutional neural networks , 2017, 2017 IEEE Intelligent Vehicles Symposium (IV).

[28]  A. Meltzoff,et al.  Like me” as a building block for understanding other minds: Bodily acts, attention, and intention. Ed. Malle, BF, L. J. Moses, and DA Baldwin , 2001 .

[29]  Mathias Perrollaz,et al.  Learning-based approach for online lane change intention prediction , 2013, 2013 IEEE Intelligent Vehicles Symposium (IV).

[30]  Tomohiro Yamamura,et al.  A Driver Behavior Recognition Method Based on a Driver Model Framework , 2000 .

[31]  K. Doya,et al.  A unifying computational framework for motor control and social interaction. , 2003, Philosophical transactions of the Royal Society of London. Series B, Biological sciences.

[32]  Siddhartha S. Srinivasa,et al.  Predicting User Intent Through Eye Gaze for Shared Autonomy , 2016, AAAI Fall Symposia.

[33]  Hema Swetha Koppula,et al.  Recurrent Neural Networks for driver activity anticipation via sensory-fusion architecture , 2015, 2016 IEEE International Conference on Robotics and Automation (ICRA).

[34]  Lina Yao,et al.  TagTrack: device-free localization and tracking using passive RFID tags , 2014, MobiQuitous.

[35]  Mohan M. Trivedi,et al.  On the Roles of Eye Gaze and Head Dynamics in Predicting Driver's Intent to Change Lanes , 2009, IEEE Transactions on Intelligent Transportation Systems.

[36]  Yuying Jiang,et al.  Driver intention recognition based on Continuous Hidden Markov Model , 2011, Proceedings 2011 International Conference on Transportation, Mechanical, and Electrical Engineering (TMEE).

[37]  Simo Särkkä,et al.  Bayesian Filtering and Smoothing , 2013, Institute of Mathematical Statistics textbooks.

[38]  Peter Stone,et al.  DIPD: Gaze-Based Intention Inference in Dynamic Environments , 2018, AAAI Workshops.

[39]  Jooyoung Park,et al.  Prediction of Driver’s Intention of Lane Change by Augmenting Sensor Information Using Machine Learning Techniques , 2017, Sensors.

[40]  Anup Doshi,et al.  Lane change intent prediction for driver assistance: On-road design and evaluation , 2011, 2011 IEEE Intelligent Vehicles Symposium (IV).

[41]  Biing-Hwang Juang,et al.  Fundamentals of speech recognition , 1993, Prentice Hall signal processing series.