Automation Expectation Mismatch: Incorrect Prediction Despite Eyes on Threat and Hands on Wheel

Objective: The aim of this study was to understand how to secure driver supervision engagement and conflict intervention performance while using highly reliable (but not perfect) automation. Background: Securing driver engagement—by mitigating irony of automation (i.e., the better the automation, the less attention drivers will pay to traffic and the system, and the less capable they will be to resume control) and by communicating system limitations to avoid mental model misconceptions—is a major challenge in the human factors literature. Method: One hundred six drivers participated in three test-track experiments in which we studied driver intervention response to conflicts after driving highly reliable but supervised automation. After 30 min, a conflict occurred wherein the lead vehicle cut out of lane to reveal a conflict object in the form of either a stationary car or a garbage bag. Results: Supervision reminders effectively maintained drivers’ eyes on path and hands on wheel. However, neither these reminders nor explicit instructions on system limitations and supervision responsibilities prevented 28% (21/76) of drivers from crashing with their eyes on the conflict object (car or bag). Conclusion: The results uncover the important role of expectation mismatches, showing that a key component of driver engagement is cognitive (understanding the need for action), rather than purely visual (looking at the threat), or having hands on wheel. Application: Automation needs to be designed either so that it does not rely on the driver or so that the driver unmistakably understands that it is an assistance system that needs an active driver to lead and share control.

[1]  Y. Benjamini,et al.  Controlling the false discovery rate: a practical and powerful approach to multiple testing , 1995 .

[2]  Charles A. Green,et al.  The “Out-of-the-Loop” concept in automated driving: proposed definition, measures and implications , 2018, Cognition, Technology & Work.

[3]  M. Bradley,et al.  Measuring emotion: the Self-Assessment Manikin and the Semantic Differential. , 1994, Journal of behavior therapy and experimental psychiatry.

[4]  T. Victor Keeping Eye and Mind on the Road , 2005 .

[5]  Nidhi Kalra,et al.  The Enemy of Good , 2017 .

[6]  P. Wiffen Change blindness , 2018, European journal of hospital pharmacy. Science and practice.

[7]  Josef F. Krems,et al.  Keep Your Scanners Peeled , 2016, Hum. Factors.

[8]  T. Åkerstedt,et al.  Validation of the Karolinska sleepiness scale against performance and EEG variables , 2006, Clinical Neurophysiology.

[9]  Christopher D. Wickens,et al.  Identifying Black Swans in NextGen: Predicting Human Performance in Off-Nominal Conditions , 2009, Hum. Factors.

[10]  Daniel J. Simons,et al.  Inattentional blindness , 2007, Scholarpedia.

[11]  E. Miller,et al.  An integrative theory of prefrontal cortex function. , 2001, Annual review of neuroscience.

[12]  George Mason Situation Awareness, Mental Workload, and Trust in Automation:Viable, Empirically Supported Cognitive Engineering Constructs , 2011 .

[13]  Raja Parasuraman,et al.  Automation in Future Air Traffic Management: Effects of Decision Aid Reliability on Controller Performance and Mental Workload , 2005, Hum. Factors.

[14]  William Payre,et al.  Fully Automated Driving , 2016, Hum. Factors.

[15]  Abuse Humans and Automation : Use , Misuse , Disuse , , 2008 .

[16]  Raja Parasuraman,et al.  Monitoring an Automated System for a Single Failure: Vigilance and Task Complexity Effects , 1996, Hum. Factors.

[17]  John D. Lee,et al.  Distraction Detection and Mitigation Through Driver Feedback , 2013 .

[18]  A. Clark Whatever next? Predictive brains, situated agents, and the future of cognitive science. , 2013, The Behavioral and brain sciences.

[19]  Neville A. Stanton,et al.  Effects of adaptive cruise control and highly automated driving on workload and situation awareness: A review of the empirical evidence , 2014 .

[20]  J. Krems,et al.  The evolution of mental model, trust and acceptance of adaptive cruise control in relation to initial information , 2013 .

[21]  Basic numbers needed to understand the traffic safety effect of Automated Cars , 2017 .

[22]  Mica R. Endsley,et al.  The Out-of-the-Loop Performance Problem and Level of Control in Automation , 1995, Hum. Factors.

[23]  Thomas B. Sheridan,et al.  Telerobotics, Automation, and Human Supervisory Control , 2003 .

[24]  Santokh Singh,et al.  Critical Reasons for Crashes Investigated in the National Motor Vehicle Crash Causation Survey , 2015 .

[25]  Nidhi Kalra,et al.  Driving to Safety , 2016 .

[26]  Masooda Bashir,et al.  Trust in Automation , 2015, Hum. Factors.

[27]  Niklas Strand,et al.  Semi-automated versus highly automated driving in critical situations caused by automation failures , 2014 .

[28]  John D. Lee,et al.  Trust in Automation: Designing for Appropriate Reliance , 2004, Hum. Factors.

[29]  Bobbie Seppelt,et al.  Potential Solutions to Human Factors Challenges in Road Vehicle Automation , 2016 .

[30]  Christopher D. Wickens,et al.  A model for types and levels of human interaction with automation , 2000, IEEE Trans. Syst. Man Cybern. Part A.

[31]  Raja Parasuraman,et al.  Complacency and Bias in Human Use of Automation: An Attentional Integration , 2010, Hum. Factors.

[32]  Annika F L Larsson,et al.  Driver usage and understanding of adaptive cruise control. , 2012, Applied ergonomics.

[33]  Marco Dozza,et al.  Analysis of Naturalistic Driving Study Data: Safer Glances, Driver Inattention, and Crash Risk , 2014 .

[34]  John D. Lee,et al.  Human Factors and Ergonomics in Automation Design , 2006 .

[35]  Mark Mulder,et al.  Introduction to the special issue on shared control , 2015, HRI 2015.

[36]  Sebastian Hergeth,et al.  Automation Trust in ConditionalAutomated Driving Systems: Approachesto Operationalization and Design , 2016 .

[37]  N. Moray,et al.  Trust in automation. Part II. Experimental studies of trust and human intervention in a process control simulation. , 1996, Ergonomics.

[38]  Johan Engström,et al.  Great expectations: a predictive processing account of automobile driving , 2018 .

[39]  Lisanne Bainbridge,et al.  Ironies of automation , 1982, Autom..

[40]  Anders Lie,et al.  Cars Are Driven on Roads, Joint Visions and Modern Technologies Stress the Need for Co-Operation , 2011 .

[41]  Huiyang Li,et al.  Human Performance Consequences of Stages and Levels of Automation , 2014, Hum. Factors.

[42]  Roger Johansson,et al.  Vision Zero – Implementing a policy for traffic safety , 2009 .

[43]  Johan Engström,et al.  Sensitivity of eye-movement measures to in-vehicle task difficulty , 2005 .

[44]  Raja Parasuraman,et al.  Humans and Automation: Use, Misuse, Disuse, Abuse , 1997, Hum. Factors.

[45]  Bobbie Danielle Seppelt,et al.  Supporting operator reliance on automation through continuous feedback , 2009 .

[46]  Nadine B. Sarter,et al.  How in the World Did We Ever Get into That Mode? Mode Error and Awareness in Supervisory Control , 1995, Hum. Factors.

[47]  Trent Victor,et al.  When Autonomous Vehicles Are Introduced on a Larger Scale in the Road Transport System: The Drive Me Project , 2017 .

[48]  Marco Dozza,et al.  A Reference Model for Driver Attention in Automation: Glance Behavior Changes During Lateral and Longitudinal Assistance , 2019, IEEE Transactions on Intelligent Transportation Systems.

[49]  Renwick E. Curry,et al.  Flight-deck automation: promises and problems , 1980 .