CoCAtt: A Cognitive-Conditioned Driver Attention Dataset

The task of driver attention prediction has drawn considerable interest among researchers in robotics and the autonomous vehicle industry. Driver attention prediction can play an instrumental role in mitigating and preventing highrisk events, like collisions and casualties. However, existing driver attention prediction models neglect the distraction state and intention of the driver, which can significantly influence how they observe their surroundings. To address these issues, we present a new driver attention dataset, CoCAtt (Cognitive-Conditioned Attention). Unlike previous driver attention datasets, CoCAtt includes per-frame annotations that describe the distraction state and intention of the driver. In addition, the attention data in our dataset is captured in both manual and autopilot modes using eyetracking devices of different resolutions. Our results demonstrate that incorporating the above two driver states into attention modeling can improve the performance of driver attention prediction. To the best of our knowledge, this work is the first to provide autopilot attention data. Furthermore, CoCAtt is currently the largest and the most diverse driver attention dataset in terms of autonomy levels, eye tracker resolutions, and driving scenarios.

[1]  Víctor Corcoba Magaña,et al.  The Effects of the Driver’s Mental State and Passenger Compartment Conditions on Driving Performance and Driving Stress , 2020, Sensors.

[2]  Ruzena Bajcsy,et al.  Optimizing interaction between humans and autonomy via information constraints on interface design , 2017, 2017 IEEE 20th International Conference on Intelligent Transportation Systems (ITSC).

[3]  Kathryn M. McMillan,et al.  N‐back working memory paradigm: A meta‐analysis of normative functional neuroimaging studies , 2005, Human brain mapping.

[4]  Ruzena Bajcsy,et al.  Semiautonomous Vehicular Control Using Driver Modeling , 2014, IEEE Transactions on Intelligent Transportation Systems.

[5]  Christiane B. Wiebel-Herboth,et al.  Human-Vehicle Cooperation on Prediction-Level: Enhancing Automated Driving with Human Foresight , 2021, 2021 IEEE Intelligent Vehicles Symposium Workshops (IV Workshops).

[6]  John D. Lee,et al.  A hybrid Bayesian Network approach to detect driver cognitive distraction , 2014 .

[7]  Antonio Torralba,et al.  Contextual guidance of eye movements and attention in real-world scenes: the role of global features in object search. , 2006, Psychological review.

[8]  Jean-Philippe Tarel,et al.  Alerting the drivers about road signs with poor visual saliency , 2009, 2009 IEEE Intelligent Vehicles Symposium.

[9]  Liang Zheng,et al.  Improving Driver Gaze Prediction With Reinforced Attention , 2021, IEEE Transactions on Multimedia.

[10]  Victoria A Banks,et al.  Is partially automated driving a bad idea? Observations from an on-road study. , 2018, Applied ergonomics.

[11]  John D. Lee,et al.  Real-Time Detection of Driver Cognitive Distraction Using Support Vector Machines , 2007, IEEE Transactions on Intelligent Transportation Systems.

[12]  Germán Ros,et al.  CARLA: An Open Urban Driving Simulator , 2017, CoRL.

[13]  Andrea Palazzi,et al.  Predicting the Driver's Focus of Attention: The DR(eye)VE Project , 2017, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[14]  David Whitney,et al.  Predicting Driver Attention in Critical Situations , 2017, ACCV.

[15]  Henrik I. Christensen,et al.  “Looking at the Right Stuff” – Guided Semantic-Gaze for Autonomous Driving , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[16]  S. Shankar Sastry,et al.  Towards trustworthy automation: User interfaces that convey internal and external awareness , 2016, 2016 IEEE 19th International Conference on Intelligent Transportation Systems (ITSC).

[17]  Quoc V. Le,et al.  CondConv: Conditionally Parameterized Convolutions for Efficient Inference , 2019, NeurIPS.

[18]  Yiannis Demiris,et al.  HammerDrive: A Task-Aware Driving Visual Attention Model , 2021, IEEE Transactions on Intelligent Transportation Systems.

[19]  Nanning Zheng,et al.  Learning to Detect a Salient Object , 2011, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[20]  Linda R. Elliott,et al.  The benefits of multimodal information: a meta-analysis comparing visual and visual-tactile feedback , 2006, ICMI '06.

[21]  William Gropp,et al.  HAL: Computer System for Scalable Deep Learning , 2020, PEARC.

[22]  Darja Topolšek,et al.  Influence of drivers’ visual and cognitive attention on their perception of changes in the traffic environment , 2019, European Transport Research Review.

[23]  Mohamed F. Marzban,et al.  The Multimodal Driver Monitoring Database: A Naturalistic Corpus to Study Driver Attention , 2021, ArXiv.

[24]  Jianru Xue,et al.  DADA-2000: Can Driving Accident be Predicted by Driver Attentionƒ Analyzed by A Benchmark , 2019, 2019 IEEE Intelligent Transportation Systems Conference (ITSC).

[25]  D. Ballard,et al.  Eye movements in natural behavior , 2005, Trends in Cognitive Sciences.

[26]  Aykut Erdem,et al.  Visual saliency estimation by nonlinearly integrating features using region covariances. , 2013, Journal of vision.

[27]  Vicente Ordonez,et al.  MEDIRL: Predicting the Visual Attention of Drivers via Maximum Entropy Deep Inverse Reinforcement Learning , 2019, 2021 IEEE/CVF International Conference on Computer Vision (ICCV).

[28]  B. Tatler,et al.  Looking and Acting: Vision and eye movements in natural behaviour , 2009 .

[29]  Patrick Le Callet,et al.  A coherent computational approach to model bottom-up visual attention , 2006, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[30]  Mohan M. Trivedi,et al.  Driver Gaze Estimation in the Real World: Overcoming the Eyeglass Challenge , 2020, 2020 IEEE Intelligent Vehicles Symposium (IV).

[31]  Ioannis Pavlidis,et al.  A multimodal dataset for various forms of distracted driving , 2017, Scientific Data.

[32]  S. W. Litvin,et al.  Looking for the Right Stuff , 2005 .

[33]  Hamidur Rahman,et al.  Driver Monitoring in the Context of Autonomous Vehicle , 2015, SCAI.

[34]  Byeongkeun Kang,et al.  High-Resolution Neural Network for Driver Visual Attention Prediction , 2020, Sensors.

[35]  Mary M Hayhoe,et al.  Visual memory and motor planning in a natural task. , 2003, Journal of vision.

[36]  S Ullman,et al.  Shifts in selective visual attention: towards the underlying neural circuitry. , 1985, Human neurobiology.

[37]  Sujitha Martin,et al.  Goal-oriented Object Importance Estimation in On-road Driving Videos , 2019, 2019 International Conference on Robotics and Automation (ICRA).

[38]  Ying Wang,et al.  Detection of Driver Cognitive Distraction: A Comparison Study of Stop-Controlled Intersection and Speed-Limited Highway , 2016, IEEE Transactions on Intelligent Transportation Systems.

[39]  Christian P. Janssen,et al.  Multitasking in Autonomous Vehicles: Ready to Go? , 2015 .

[40]  D. Ballard,et al.  Modeling Task Control of Eye Movements , 2014, Current Biology.

[41]  Christof Koch,et al.  A Model of Saliency-Based Visual Attention for Rapid Scene Analysis , 2009 .