Explainable Reinforcement Learning in Human-Robot Teams: The Impact of Decision-Tree Explanations on Transparency

Understanding the decisions of AI-driven systems and the rationale behind such decisions is key to the success of the human-robot team. However, the complexity and the "black-box" nature of many AI algorithms create a barrier for establishing such understanding within their human counterparts. Reinforcement Learning (RL), a machine-learning algorithm based on the simple idea of action-reward mappings, has a rich quantitative representation and a complex iterative reasoning process that present a significant obstacle to human understanding of, for example, how value functions are constructed, how the algorithms update the value functions, and how such updates impact the action/policy chosen by the robot. In this paper, we discuss our work to address this challenge by developing a decision-tree based explainable model for RL to make a robot’s decision-making process more transparent. Set in a human-robot virtual teaming testbed, we conducted a study to assess the impact of the explanations, generated using decision trees, on building transparency, calibrating trust, and improving the overall human-robot team’s performance. We discuss the design of the explainable model and the positive impact of the explanations on outcome measures.

[1]  Ranu Sewada,et al.  Explainable Artificial Intelligence (XAI) , 2023, international journal of food and nutritional sciences.

[2]  Natalia Díaz Rodríguez,et al.  Explainability in Deep Reinforcement Learning , 2020, Knowl. Based Syst..

[3]  Bradley Hayes,et al.  Explanation-Based Reward Coaching to Improve Human Performance via Reinforcement Learning , 2019, 2019 14th ACM/IEEE International Conference on Human-Robot Interaction (HRI).

[4]  Gang Wang,et al.  LEMNA: Explaining Deep Learning based Security Applications , 2018, CCS.

[5]  Amina Adadi,et al.  Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI) , 2018, IEEE Access.

[6]  Adnan Darwiche,et al.  A Symbolic Approach to Explaining Bayesian Network Classifiers , 2018, IJCAI.

[7]  Ning Wang,et al.  Is It My Looks? Or Something I Said? The Impact of Explanations, Embodiment, and Expectations on Trust and Performance in Human-Robot Teams , 2018, PERSUASIVE.

[8]  R. M. Taylor,et al.  Situational Awareness Rating Technique (Sart): The Development of a Tool for Aircrew Systems Design , 2017 .

[9]  Ning Wang,et al.  The Impact of POMDP-Generated Explanations on Trust and Performance in Human-Robot Teams , 2016, AAMAS.

[10]  Marco Tulio Ribeiro,et al.  “Why Should I Trust You?”: Explaining the Predictions of Any Classifier , 2016, NAACL.

[11]  Gregory Dudek,et al.  OPTIMo: Online Probabilistic Trust Inference Model for Asymmetric Human-Robot Collaborations , 2015, 2015 10th ACM/IEEE International Conference on Human-Robot Interaction (HRI).

[12]  Nicholas Mattei,et al.  A Natural Language Argumentation Interface for Explanation Generation in Markov Decision Processes , 2011, ExaCt.

[13]  Raja Parasuraman,et al.  Complacency and Bias in Human Use of Automation: An Attentional Integration , 2010, Hum. Factors.

[14]  Pascal Poupart,et al.  Minimal Sufficient Explanations for Factored Markov Decision Processes , 2009, ICAPS.

[15]  K. Dautenhahn,et al.  The Negative Attitudes Towards Robots Scale and reactions to robot behaviour in a live Human-Robot Interaction study , 2009 .

[16]  Eric T. Bradlow,et al.  Promises and Lies: Restoring Violated Trust , 2004 .

[17]  Regina A. Pomranky,et al.  The role of trust in automation reliance , 2003, Int. J. Hum. Comput. Stud..

[18]  Charles J. Kacmar,et al.  Developing and Validating Trust Measures for e-Commerce: An Integrative Typology , 2002, Inf. Syst. Res..

[19]  V. Greco,et al.  Coping with uncertainty: the construction and validation of a new measure , 2001 .

[20]  John Riedl,et al.  Explaining collaborative filtering recommendations , 2000, CSCW '00.

[21]  Raja Parasuraman,et al.  Humans and Automation: Use, Misuse, Disuse, Abuse , 1997, Hum. Factors.

[22]  Andrew W. Moore,et al.  Reinforcement Learning: A Survey , 1996, J. Artif. Intell. Res..

[23]  J. H. Davis,et al.  An Integrative Model Of Organizational Trust , 1995 .

[24]  Johanna D. Moore,et al.  Explanation in second generation expert systems , 1993 .

[25]  N Moray,et al.  Trust, control strategies and allocation of function in human-machine systems. , 1992, Ergonomics.

[26]  Johanna D. Moore,et al.  Explanations in knowledge systems: design for explainable expert systems , 1991, IEEE Expert.

[27]  L. Hendrickx,et al.  Relative importance of scenario information and frequency information in the judgment of risk , 1989 .

[28]  Bonnie M. Muir,et al.  Trust Between Humans and Machines, and the Design of Decision Aids , 1987, Int. J. Man Mach. Stud..

[29]  Or Biran,et al.  Explanation and Justification in Machine Learning : A Survey Or , 2017 .

[30]  David V. Pynadath,et al.  Building Trust in a Human-Robot Team with Automatically Generated Explanations , 2015 .

[31]  J. M. Ross Moderators Of Trust And Reliance Across Multiple Decision Aids , 2008 .

[32]  F. Elizalde,et al.  Policy Explanation in Factored Markov Decision Processes , 2008 .

[33]  Milind Tambe,et al.  Electric Elves: What Went Wrong and Why , 2006, AAAI Spring Symposium: What Went Wrong and Why: Lessons from AI Research and Applications.

[34]  J. R. Quinlan Induction of decision trees , 2004, Machine Learning.

[35]  Richard S. Sutton,et al.  Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.

[36]  S. Hart,et al.  Development of NASA-TLX (Task Load Index): Results of Empirical and Theoretical Research , 1988 .