How to Interact with a Fully Autonomous Vehicle: Naturalistic Ways for Drivers to Intervene in the Vehicle System While Performing Non-Driving Related Tasks

Autonomous vehicle technology increasingly allows drivers to turn their primary attention to secondary tasks (e.g., eating or working). This dramatic behavior change thus requires new input modalities to support driver–vehicle interaction, which must match the driver’s in-vehicle activities and the interaction situation. Prior studies that addressed this question did not consider how acceptance for inputs was affected by the physical and cognitive levels experienced by drivers engaged in Non-driving Related Tasks (NDRTs) or how their acceptance varies according to the interaction situation. This study investigates naturalistic interactions with a fully autonomous vehicle system in different intervention scenarios while drivers perform NDRTs. We presented an online methodology to 360 participants showing four NDRTs with different physical and cognitive engagement levels, and tested the six most common intervention scenarios (24 cases). Participants evaluated our proposed seven natural input interactions for each case: touch, voice, hand gesture, and their combinations. Results show that NDRTs influence the driver’s input interaction more than intervention scenario categories. In contrast, variation of physical load has more influence on input selection than variation of cognitive load. We also present a decision-making model of driver preferences to determine the most natural inputs and help User Experience designers better meet drivers’ needs.

[1]  Hyang Sook Kim,et al.  Deriving future user experiences in autonomous vehicle , 2015, AutomotiveUI.

[2]  Martin Baumann,et al.  Cooperative Overtaking: Overcoming Automated Vehicles' Obstructed Sensor Range via Driver Help , 2019, AutomotiveUI.

[3]  Alex Fridman,et al.  Behavioral Impact of Drivers' Roles in Automated Driving , 2016, AutomotiveUI.

[4]  Tom Gross,et al.  I See Your Point: Integrating Gaze to Enhance Pointing Gesture Accuracy While Driving , 2018, AutomotiveUI.

[5]  H. Simon,et al.  Rational choice and the structure of the environment. , 1956, Psychological review.

[6]  Christian A. Müller,et al.  Multimodal Input in the Car, Today and Tomorrow , 2011, IEEE MultiMedia.

[7]  Michael D Fetters,et al.  Video Elicitation Interviews: A Qualitative Research Method for Investigating Physician-Patient Interactions , 2012, The Annals of Family Medicine.

[8]  M. König,et al.  Users’ resistance towards radical innovations: The case of the self-driving car , 2017 .

[9]  Stefan Schneegaß,et al.  User-Defined Voice and Mid-Air Gesture Commands for Maneuver-based Interventions in Automated Vehicles , 2019, MuC.

[10]  Henrik Detjen,et al.  Maneuver-based Driving for Intervention in Autonomous Cars , 2020, ArXiv.

[11]  Anind K. Dey,et al.  Simulated augmented reality windshield display as a cognitive mapping aid for elder driver navigation , 2009, CHI.

[12]  Riender Happee,et al.  Public opinion on automated driving: results of an international questionnaire among 5000 respondents , 2015 .

[13]  Philipp Wintersberger,et al.  Why do you like to drive automated?: a context-dependent analysis of highly automated driving to elaborate requirements for intelligent user interfaces , 2019, IUI.

[14]  Michael Weber,et al.  Towards Cooperative Driving: Involving the Driver in an Autonomous Vehicle's Decision Making , 2016, AutomotiveUI.

[15]  Alex Fridman,et al.  To Walk or Not to Walk: Crowdsourced Assessment of External Vehicle-to-Pedestrian Displays , 2017, ArXiv.

[16]  Pierluigi Pisu,et al.  VoGe: A Voice and Gesture System for Interacting with Autonomous Cars , 2017 .

[17]  Manfred Tscheligi,et al.  Towards Autonomous Cars: The Effect of Autonomy Levels on Acceptance and User Experience , 2014, AutomotiveUI.

[18]  Jacques M. B. Terken,et al.  Compatibility between Trust and Non-Driving Related Tasks in UI Design for Highly and Fully Automated Driving , 2016, AutomotiveUI.

[19]  Klaus Bengler,et al.  From HMI to HMIs: Towards an HMI Framework for Automated Driving , 2020, Inf..

[20]  Pierre-Majorique Léger,et al.  When Should I Use my Active Workstation? The impact of Physical Demand and Task Difficulty on IT Users’ Perception and Performance , 2017 .

[21]  M. L. Plume,et al.  SPSS (Statistical Package for the Social Sciences) , 2002, Encyclopedia of Information Systems.

[22]  Johannes Kraus,et al.  Calibration of Trust Expectancies in Conditionally Automated Driving by Brand, Reliability Information and Introductionary Videos: An Online Study , 2018, AutomotiveUI.

[23]  Marc Erich Latoschik,et al.  “Stop over there”: natural gesture and speech interaction for non-critical spontaneous intervention in autonomous driving , 2017, ICMI.

[24]  Seungjun Kim,et al.  A Cascaded Multimodal Natural User Interface to Reduce Driver Distraction , 2020, IEEE Access.

[25]  Harsh Sanghavi,et al.  Exploring User Needs and Design Requirements in Fully Automated Vehicles , 2020, CHI Extended Abstracts.

[26]  Gary Klein,et al.  Naturalistic Decision Making , 2008, Hum. Factors.

[27]  Philipp Wintersberger,et al.  Driving Hotzenplotz: A Hybrid Interface for Vehicle Control Aiming to Maximize Pleasure in Highway Driving , 2017, AutomotiveUI.

[28]  Neville A. Stanton,et al.  Turing in the driver's seat: Can people distinguish between automated and manually driven vehicles? , 2020, Human Factors and Ergonomics in Manufacturing & Service Industries.

[29]  V. Braun,et al.  The online survey as a qualitative research tool , 2020, International Journal of Social Research Methodology.

[30]  M. Endsley The role of situation awareness in naturalistic decision making , 1997 .

[31]  Radu-Daniel Vatavu,et al.  A Synopsis of Input Modalities for In-Vehicle Infotainment and Consumption of Interactive Media , 2020, IMX.

[32]  Mohammed Elhenawy,et al.  Self-Interruptions of Non-Driving Related Tasks in Automated Vehicles: Mobile vs Head-Up Display , 2020, CHI.

[33]  Shigeki Sugano,et al.  Analysis of individual driving experience in autonomous and human-driven vehicles using a driving simulator , 2015, 2015 IEEE International Conference on Advanced Intelligent Mechatronics (AIM).

[34]  Taxonomy and definitions for terms related to driving automation systems for on-road motor vehicles , 2022 .

[35]  Sha Cao,et al.  Voice User Interface Interaction Design Research Based on User Mental Model in Autonomous Vehicle , 2018, HCI.

[36]  Shigeki Sugano,et al.  A hand gesture based driver-vehicle interface to control lateral and longitudinal motions of an autonomous vehicle , 2016, 2016 IEEE International Conference on Systems, Man, and Cybernetics (SMC).

[37]  Paul A. Jennings,et al.  Evaluating How Interfaces Influence the User Interaction with Fully Autonomous Vehicles , 2018, AutomotiveUI.

[38]  Chao Wang A framework of the non-critical spontaneous intervention in highly automated driving scenarios , 2019, AutomotiveUI.

[39]  Wendy Ju,et al.  Designing robots with movement in mind , 2014, Journal of Human-Robot Interaction.

[40]  David Sirkin,et al.  What a Driver Wants: User Preferences in Semi-Autonomous Vehicle Decision-Making , 2020, CHI.

[41]  Michael Weber,et al.  Touch Screen Maneuver Approval Mechanisms for Highly Automated Vehicles: A First Evaluation , 2017, AutomotiveUI.

[42]  Natasha Merat,et al.  Highly Automated Driving, Secondary Task Performance, and Driver State , 2012, Hum. Factors.

[43]  Orestis Georgiou,et al.  Designing Mid-Air Haptic Gesture Controlled User Interfaces for Cars , 2020, Proc. ACM Hum. Comput. Interact..

[44]  Myra Blanco,et al.  The impact of secondary task cognitive processing demand on driving performance. , 2006, Accident; analysis and prevention.

[45]  Gunnar Stevens,et al.  Using Time and Space Efficiently in Driverless Cars: Findings of a Co-Design Study , 2019, CHI.

[46]  Stewart A. Birrell,et al.  The influence of system transparency on trust: Evaluating interfaces in a highly automated vehicle , 2020, Transportation Research Part F: Traffic Psychology and Behaviour.

[47]  Hermann Winner,et al.  Conduct-by-Wire : following a new paradigm for driving into the future , 2006 .

[48]  Carol A. C. Flannagan,et al.  ATD positioning based on driver posture and position , 1998 .

[49]  Michael Feld,et al.  Combining Speech, Gaze, and Micro-gestures for the Multimodal Control of In-Car Functions , 2016, 2016 12th International Conference on Intelligent Environments (IE).

[50]  Geehyuk Lee,et al.  Voice+Tactile: Augmenting In-vehicle Voice User Interface with Tactile Touchpad Interaction , 2020, CHI.

[51]  Ralph Bruder,et al.  How to conduct a car? A design example for maneuver based driver-vehicle interaction , 2010, 2010 IEEE Intelligent Vehicles Symposium.

[52]  Tobias Hecht,et al.  Non-driving Related Activities in Automated Driving - An Online Survey Investigating User Needs , 2019, IHSED.

[53]  Klaus Bengler,et al.  Comparison of Methods to Evaluate the Influence of an Automated Vehicle's Driving Behavior on Pedestrians: Wizard of Oz, Virtual Reality, and Video , 2020, Inf..

[54]  Yong Gu Ji,et al.  Non-driving-related tasks, workload, and takeover performance in highly automated driving contexts , 2019, Transportation Research Part F: Traffic Psychology and Behaviour.

[55]  Aniket Kittur,et al.  Crowdsourcing user studies with Mechanical Turk , 2008, CHI.

[56]  Christoph Bartneck,et al.  Can we control it? Autonomous robots threaten human identity, uniqueness, safety, and resources , 2017, Int. J. Hum. Comput. Stud..

[57]  Michael Greatorex,et al.  Statistical Package for the Social Sciences , 2015 .

[58]  David R. Large,et al.  Design implications of drivers’ engagement with secondary activities during highly-automated driving – a longitudinal simulator study , 2017 .

[59]  Simon Thompson,et al.  Information Expectations in Highly and Fully Automated Vehicles , 2017 .

[60]  Bastian Pfleging,et al.  How to Increase Automated Vehicles’ Acceptance through In-Vehicle Interaction Design: A Review , 2021, Int. J. Hum. Comput. Interact..

[61]  Daisong Guan,et al.  A classification framework based on driver's operations of in-car interaction , 2019, AutomotiveUI.

[62]  Bastian Pfleging,et al.  Investigating user needs for non-driving-related activities during automated driving , 2016, MUM.

[63]  Tobias Hecht,et al.  What Do You Do? An Analysis of Non-driving Related Activities During a 60 Minutes Conditionally Automated Highway Drive , 2019, IHIET.

[64]  Seungjun Kim,et al.  Gaze-Head Input: Examining Potential Interaction with Immediate Experience Sampling in an Autonomous Vehicle , 2020, Applied Sciences.

[65]  A. Ghasemi,et al.  Normality Tests for Statistical Analysis: A Guide for Non-Statisticians , 2012, International journal of endocrinology and metabolism.

[66]  James D. Abbey,et al.  Attention by design: Using attention checks to detect inattentive respondents and improve data quality , 2017 .