Detecting Human Trust Calibration in Automation: A Deep Learning Approach

INTRODUCTION Complicated automated systems are misused and disused by humans, even though the systems provide probable advantages by aiding them in satisfying their tasks (Lee & See, 2004). The misuse (over-reliance) or disuse (under-reliance) of automation was considered as trust of people to automation. For instance, the automated systems tend to be used by people who trust the system than those who do not trust that. More importantly, misuse and disuse of automation often occur when trust of operator in automation is poorly adjusted (Merrit & Ilgen, 2008). Trust calibration is the case by which the operator’s trust of automated system toward automation is adjusted based on the physical system performance and capabilities (McGuirl & Sarter, 2006). Several factors that influence trust calibration have been found, including reliability (Dzindolet et al., 2003), automation errors (Wiegmann et al., 2001), individual differences in the causal attribution of automation errors (Pop et al., 2015), and task difficulty (Madhavan et al., 2006). However, there is still a lack of studies that monitor and detect the trust adjustment of an operator in an automated system by a general quantitative model during their task. There are several reasons for this. First, studies have revealed that changed trust of an operator towards automation can be found (e.g., Dzindolet et al., 2003; Madhavan et al., 2006; Pop et al., 2015; Wiegmann et al., 2001), the adjusted trust was decided only after the task is finished. Second, several models related with prediction of human trust have been suggested mainly using discrete, self-reported behavioral measurements (e.g., Jonker & Treur, 1999) and continuous cognitive signals (e.g., Riedl & Javor, 2012). Neural correlates of trust between humans (e.g., Krueger et al., 2007) and between human and machine (e.g., Desmet et al., 2014; Goodyear et al., 2017) were also studied. But only a few studies investigated neural correlates of trust calibration in automation, in particular using electroencephalograms (EEGs) that have better temporal dynamics than other neuroimaging methods (e.g., fMRI) (e.g., de Visser et al., 2018). Finally, Deep Learning (DL) has been widely applied in many challenging domains owing to its higher success in many applications than existing machine learning. For example, Convolutional Neural Networks (CNN) has been showing promising performance in neuroergonomics fields (e.g., mental workload (Zhang et al., 2017), motor imagery (Yang et al., 2015)). However, there was almost no research on developing an EEG-based DL method that can detect an operator’s trust calibration in automation dynamically. One of the main goals of the present study was to propose a CNN-based framework to detect human trust calibration in automation using image features of EEG preserving spectral and spatial information. The effectiveness of the proposed framework was validated against a Multi-Layer Perceptron (MLP) and Deep Neural Network (DNN).

[1]  Nadine B. Sarter,et al.  Supporting Trust Calibration and the Effective Use of Decision Aids by Presenting Dynamic System Confidence Information , 2006, Hum. Factors.

[2]  Francis T. Durso,et al.  Individual Differences in the Calibration of Trust in Automation , 2015, Hum. Factors.

[3]  N. Birbaumer,et al.  BCI2000: a general-purpose brain-computer interface (BCI) system , 2004, IEEE Transactions on Biomedical Engineering.

[4]  Regina A. Pomranky,et al.  The role of trust in automation reliance , 2003, Int. J. Hum. Comput. Stud..

[5]  Craig G. McDonald,et al.  Learning From the Slips of Others: Neural Correlates of Trust in Automated Agents , 2018, Front. Hum. Neurosci..

[6]  Catholijn M. Jonker,et al.  Formal Analysis of Models for the Dynamics of Trust Based on Experiences , 1999, MAAMAW.

[7]  Dylan D. Schmorrow,et al.  DARPA's Augmented Cognition Program-tomorrow's human computer interaction from vision to reality: building cognitively aware computational systems , 2002, Proceedings of the IEEE 7th Conference on Human Factors and Power Plants.

[8]  Guigang Zhang,et al.  Deep Learning , 2016, Int. J. Semantic Comput..

[9]  A. M. Rich,et al.  Automated diagnostic aids: The effects of aid reliability on users' trust and reliance , 2001 .

[10]  Marcel Brass,et al.  How social is error observation? The neural mechanisms underlying the observation of human and machine errors. , 2014, Social cognitive and affective neuroscience.

[11]  R. Riedl,et al.  The Biology of Trust: Integrating Evidence From Genetics, Endocrinology, and Functional Brain Imaging , 2012 .

[12]  Yan Liu,et al.  Data Augmentation for EEG-Based Emotion Recognition with Deep Convolutional Neural Networks , 2018, MMM.

[13]  G. Gomez Automatic Artifact Removal ( AAR ) toolbox v 1 . 3 ( Release 09 . 12 . 2007 ) for MATLAB , 2007 .

[14]  R. Parasuraman,et al.  An fMRI and effective connectivity study investigating miss errors during advice utilization from human and machine agents , 2017, Social neuroscience.

[15]  Daniel R. Ilgen,et al.  Not All Trust Is Created Equal: Dispositional and History-Based Trust in Human-Automation Interactions , 2008, Hum. Factors.

[16]  Cuntai Guan,et al.  On the use of convolutional neural networks and augmented CSP features for multi-class motor imagery of EEG signals classification , 2015, 2015 37th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC).

[17]  Douglas A. Wiegmann,et al.  Automation Failures on Tasks Easily Performed by Operators Undermines Trust in Automated Aids , 2003 .

[18]  John D. Lee,et al.  Trust in Automation: Designing for Appropriate Reliance , 2004, Hum. Factors.

[19]  Rubin Wang,et al.  Pattern Recognition of Momentary Mental Workload Based on Multi-Channel Electrophysiological Data and Ensemble Convolutional Neural Networks , 2017, Front. Neurosci..

[20]  H. Lüders,et al.  American Electroencephalographic Society Guidelines for Standard Electrode Position Nomenclature , 1991, Journal of clinical neurophysiology : official publication of the American Electroencephalographic Society.

[21]  N. Kriegeskorte,et al.  Neural correlates of trust , 2007, Proceedings of the National Academy of Sciences.

[22]  Arnaud Delorme,et al.  EEGLAB, SIFT, NFT, BCILAB, and ERICA: New Tools for Advanced EEG Processing , 2011, Comput. Intell. Neurosci..

[23]  A. Mognon,et al.  ADJUST: An automatic EEG artifact detector based on the joint use of spatial and temporal features. , 2011, Psychophysiology.

[24]  W. D. Miller,et al.  The U.S. Air Force-Developed Adaptation of the Multi-Attribute Task Battery for the Assessment of Human Operator Workload and Strategic Behavior , 2010 .