Computational Modeling of the Dynamics of Human Trust During Human–Machine Interactions

We developed an experiment to elicit human trust dynamics in human–machine interaction contexts and established a quantitative model of human trust behavior with respect to these contexts. The proposed model describes human trust level as a function of experience, cumulative trust, and expectation bias. We estimated the model parameters using human subject data collected from two experiments. Experiment 1 was designed to excite human trust dynamics using multiple transitions in trust level. Five hundred and eighty-one individuals participated in this experiment. Experiment 2 was an augmentation of Experiment 1 designed to study and incorporate the effects of misses and false alarms in the general model. Three hundred and thirty-three individuals participated in Experiment 2. Beyond considering the dynamics of human trust in automation, this model also characterizes the effects of demographic factors on human trust. In particular, our results show that the effects of national culture and gender on trust are significant. For example, U.S. participants showed a lower trust level and were more sensitive to misses as compared with Indian participants. The resulting trust model is intended for the development of autonomous systems that can respond to changes in human trust level in real time.

[1]  Rachel T. A. Croson,et al.  Gender and Culture: International Experimental Evidence from Trust Games , 1999 .

[2]  Raja Parasuraman,et al.  Humans and Automation: Use, Misuse, Disuse, Abuse , 1997, Hum. Factors.

[3]  Joachim Meyer,et al.  Effects of Warning Validity and Proximity on Responses to Warnings , 2001, Hum. Factors.

[4]  Klaus Bengler,et al.  Trust in Automation as a Matter of Media Influence and Experi-ence of Automated Vehicles , 2016 .

[5]  Juergen Sauer,et al.  Experience of automation failures in training: effects on trust, automation bias, complacency and performance , 2016, Ergonomics.

[6]  Neera Jain,et al.  Dynamic modeling of trust in human-machine interactions , 2017, 2017 American Control Conference (ACC).

[7]  U. Fischbacher,et al.  Decomposing trust: Explaining national trust differences , 2008 .

[8]  Douglas A. Wiegmann,et al.  Automation Failures on Tasks Easily Performed by Operators Undermines Trust in Automated Aids , 2003 .

[9]  Mark Hoogendoorn,et al.  Modelling biased human trust dynamics , 2013, Web Intell. Agent Syst..

[10]  Regina A. Pomranky,et al.  The role of trust in automation reliance , 2003, Int. J. Hum. Comput. Stud..

[11]  Mark Hoogendoorn,et al.  Design and validation of a relative trust model , 2014, Knowl. Based Syst..

[12]  Lilly Irani,et al.  Amazon Mechanical Turk , 2018, Advances in Intelligent Systems and Computing.

[13]  Holly A. H. Handley,et al.  Trust and the Compliance–Reliance Paradigm: The Effects of Risk, Error Bias, and Reliability on Trust and Dependence , 2017, Hum. Factors.

[14]  D. Pfeffermann The Role of Sampling Weights when Modeling Survey Data , 1993 .

[15]  Bonnie M. Muir,et al.  Trust Between Humans and Machines, and the Design of Decision Aids , 1987, Int. J. Man Mach. Stud..

[16]  Christopher D. Wickens,et al.  Automation Reliability in Unmanned Aerial Vehicle Control: A Reliance-Compliance Model of Automation Dependence in High Workload , 2006, Hum. Factors.

[17]  Raja Parasuraman,et al.  Human-Automation Interaction , 2005 .

[18]  Matt Richtel,et al.  Google’s Driverless Cars Run into Problem: Cars with Drivers , 2015 .

[19]  Gert Jan Hofstede,et al.  Cultures and Organizations: Software of the Mind, 3rd ed. , 2010 .

[20]  Hamed Saeidi,et al.  Modeling and Control of Trust in Human and Robot Collaborative Manufacturing , 2014, AAAI Spring Symposia.

[21]  David Gefen,et al.  On the Need to Include National Culture as a Central Issue in E-Commerce Trust Beliefs , 2006, J. Glob. Inf. Manag..

[22]  Joseph P. Cannon,et al.  Understanding the Influence of National Culture on the Development of Trust , 1998 .

[23]  Jessie Y.C. Chen,et al.  Effects of imperfect automation and individual differences on concurrent performance of military and robotics tasks in a simulated multitasking environment , 2009, Ergonomics.

[24]  Detmar W. Straub,et al.  Examining Trust in Information Technology Artifacts: The Effects of System Quality and Culture , 2008, J. Manag. Inf. Syst..

[25]  Bernard C. Y. Tan,et al.  Dimensions of national culture and corporate adoption of IT infrastructure , 2001, IEEE Trans. Engineering Management.

[26]  C. Carver,et al.  Attention and Self-Regulation: A Control-Theory Approach to Human Behavior , 1981 .

[27]  Christopher D. Wickens,et al.  False Alerts in Air Traffic Control Conflict Alerting System: Is There a “Cry Wolf” Effect? , 2009, Hum. Factors.

[28]  Katia P. Sycara,et al.  Influence of cultural factors in dynamic trust in automation , 2016, 2016 IEEE International Conference on Systems, Man, and Cybernetics (SMC).

[29]  J. H. Davis,et al.  An Integrative Model of Organizational Trust: Past, Present, and Future , 2007 .

[30]  Martin Gonzalez-Rodriguez,et al.  Web Design Attributes in Building User Trust, Satisfaction, and Loyalty for a High Uncertainty Avoidance Culture , 2017, IEEE Transactions on Human-Machine Systems.

[31]  Ernesto A. Bustamante,et al.  Effects of False-Alarm vs. Miss-Prone Automation and Likelihood Alarm Technology on Trust, Reliance, and Compliance in a Miss-Prone Task , 2010 .

[32]  Masooda Bashir,et al.  Trust in Automation , 2015, Hum. Factors.

[33]  Nash Stanton,et al.  The Effects of System Technology and Probability Type on Trust, Compliance, and Reliance , 2009 .

[34]  S Lewandowsky,et al.  The dynamics of trust: comparing humans to automation. , 2000, Journal of experimental psychology. Applied.

[35]  Jessie Y. C. Chen,et al.  Human–Agent Teaming for Multirobot Control: A Review of Human Factors Issues , 2014, IEEE Transactions on Human-Machine Systems.

[36]  Joyce E. Berg,et al.  Trust, Reciprocity, and Social History , 1995 .

[37]  T. Nomura,et al.  Exploring effects of educational backgrounds and gender in human-robot interaction , 2011, 2011 International Conference on User Science and Engineering (i-USEr ).

[38]  Catholijn M. Jonker,et al.  Formal Analysis of Models for the Dynamics of Trust Based on Experiences , 1999, MAAMAW.

[39]  N Moray,et al.  Trust, control strategies and allocation of function in human-machine systems. , 1992, Ergonomics.

[40]  Tatsuya Nomura,et al.  Prediction of Human Behavior in Human--Robot Interaction Using Psychological Scales for Anxiety and Negative Attitudes Toward Robots , 2008, IEEE Transactions on Robotics.

[41]  Sophocles J. Orfanidis,et al.  Introduction to signal processing , 1995 .

[42]  H. Vincent Poor,et al.  An Introduction to Signal Detection and Estimation , 1994, Springer Texts in Electrical Engineering.

[43]  TerryAnn Glandon,et al.  Framing, decision-aid systems, and culture: Exploring influences on fraud investigations , 2012, Int. J. Account. Inf. Syst..

[44]  Christopher D. Wickens,et al.  On the Independence of Compliance and Reliance: Are Automation False Alarms Worse Than Misses? , 2007, Hum. Factors.

[45]  Mia Hubert,et al.  Robust statistics for outlier detection , 2011, WIREs Data Mining Knowl. Discov..

[46]  Gary Klein,et al.  Naturalistic Decision Making , 2008, Hum. Factors.

[47]  Cynthia Breazeal,et al.  Persuasive Robotics: The influence of robot gender on human behavior , 2009, 2009 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[48]  Jordi Sabater-Mir,et al.  Review on Computational Trust and Reputation Models , 2005, Artificial Intelligence Review.

[49]  James P. Bliss,et al.  Collective Mistrust of Alarms , 2002 .

[50]  G. Hofstede,et al.  Culture′s Consequences: International Differences in Work-Related Values , 1980 .

[51]  Brent A. Scott,et al.  Trust, trustworthiness, and trust propensity: a meta-analytic test of their unique relationships with risk taking and job performance. , 2007, The Journal of applied psychology.

[52]  Catholijn M. Jonker,et al.  Human Experiments in Trust Dynamics , 2004, iTrust.

[53]  Venkatesh,et al.  A Longitudinal Field Investigation of Gender Differences in Individual Technology Adoption Decision-Making Processes. , 2000, Organizational behavior and human decision processes.

[54]  Mark Hoogendoorn,et al.  Modeling and Validation of Biased Human Trust , 2011, 2011 IEEE/WIC/ACM International Conferences on Web Intelligence and Intelligent Agent Technology.

[55]  Timothy G. Rosser,et al.  Passengers from India and the United States Have Differential Opinions about Autonomous Auto-Pilots for Commercial Flights , 2014 .

[56]  Fang-Wu Tung,et al.  Influence of Gender and Age on the Attitudes of Children towards Humanoid Robots , 2011, HCI.

[57]  Ananish Chaudhuri,et al.  Gender Differences in Trust and Reciprocity , 2003 .

[58]  S. Breznitz Cry Wolf: The Psychology of False Alarms , 1984 .

[59]  John D. Lee,et al.  Trust, self-confidence, and operators' adaptation to automation , 1994, Int. J. Hum. Comput. Stud..

[60]  John D. Lee,et al.  Trust in Automation: Designing for Appropriate Reliance , 2004 .

[61]  Xin Li,et al.  Using attitude and social influence to develop an extended trust model for information systems , 2006, DATB.

[62]  G. Hofstede Culture′s Consequences: Comparing Values, Behaviors, Institutions and Organizations Across Nations , 2001 .

[63]  Arthur D. Fisk,et al.  Type of Automation Failure: The Effects on Trust and Reliance in Automation , 2004 .

[64]  Karen Zita Haigh,et al.  A Model of Trust and Reliance of Automation Technology for Older Users , 2005, AAAI Fall Symposium: Caring Machines.