Explainable robotic systems: understanding goal-driven actions in a reinforcement learning scenario

Robotic systems are more present in our society everyday. In human-robot environments, it is crucial that end-users may correctly understand their robotic team-partners, in order to collaboratively complete a task. To increase action understanding, users demand more explainability about the decisions by the robot in particular situations. Recently, explainable robotic systems have emerged as an alternative focused not only on completing a task satisfactorily, but also in justifying, in a human-like manner, the reasons that lead to making a decision. In reinforcement learning scenarios, a great effort has been focused on providing explanations using data-driven approaches, particularly from the visual input modality in deep learning-based systems. In this work, we focus on the decision-making process of a reinforcement learning agent performing a simple navigation task in a robotic scenario. As a way to explain the goal-driven robot's actions, we use the probability of success computed by three different proposed approaches: memory-based, learning-based, and introspection-based. The difference between these approaches is the amount of memory required to compute or estimate the probability of success as well as the kind of reinforcement learning representation where they could be used. In this regard, we use the memory-based approach as a baseline since it is obtained directly from the agent's observations. When comparing the learning-based and the introspection-based approaches to this baseline, both are found to be suitable alternatives to compute the probability of success, obtaining high levels of similarity when compared using both the Pearson's correlation and the mean squared error.

[1]  E. Vincent Cross,et al.  Explaining robot actions , 2012, 2012 7th ACM/IEEE International Conference on Human-Robot Interaction (HRI).

[2]  Anind K. Dey,et al.  Why and why not explanations improve the intelligibility of context-aware intelligent systems , 2009, CHI.

[3]  Jan Peters,et al.  Reinforcement learning in robotics: A survey , 2013, Int. J. Robotics Res..

[4]  Stephanie Rosenthal,et al.  Visual Explanations for Convolutional Neural Networks via Input Resampling , 2017, ArXiv.

[5]  Stefan Wermter,et al.  Improving interactive reinforcement learning: What makes a good teacher? , 2018, Connect. Sci..

[6]  Alan Fern,et al.  Strategic Tasks for Explainable Reinforcement Learning , 2019, AAAI.

[7]  Richard S. Sutton,et al.  Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.

[8]  Amina Adadi,et al.  Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI) , 2018, IEEE Access.

[9]  Davide Calvaresi,et al.  Explainable Agents and Robots: Results from a Systematic Literature Review , 2019, AAMAS.

[10]  Eric Yeh,et al.  Interestingness Elements for Explainable Reinforcement Learning through Introspection , 2019, IUI Workshops.

[11]  Kevin Li,et al.  Evaluating Effects of User Experience and System Transparency on Trust in Automation , 2017, 2017 12th ACM/IEEE International Conference on Human-Robot Interaction (HRI.

[12]  Richard Socher,et al.  Hierarchical and Interpretable Skill Acquisition in Multi-task Reinforcement Learning , 2017, ICLR.

[13]  Maria Fox,et al.  Explainable Planning , 2017, ArXiv.

[14]  E. Thorndike Animal Intelligence; Experimental Studies , 2009 .

[15]  Avi Rosenfeld,et al.  Explainability in human–agent systems , 2019, Autonomous Agents and Multi-Agent Systems.

[16]  Andrew Y. Ng,et al.  Policy Invariance Under Reward Transformations: Theory and Application to Reward Shaping , 1999, ICML.

[17]  Tom Schaul,et al.  StarCraft II: A New Challenge for Reinforcement Learning , 2017, ArXiv.

[18]  Gonzalo Acuña Leiva,et al.  Indirect Training with Error Backpropagation in Gray-Box Neural Model: Application to a Chemical Process , 2010, 2010 XXIX International Conference of the Chilean Computer Science Society.

[19]  Jonathan Dodge,et al.  Visualizing and Understanding Atari Agents , 2017, ICML.

[20]  Dawn M. Tilbury,et al.  Explanations and Expectations: Trust Building in Automated Vehicles , 2018, HRI.

[21]  Maartje M. A. de Graaf,et al.  How People Explain Action (and Autonomous Intelligent Systems Should Too) , 2017, AAAI Fall Symposia.

[22]  Katia P. Sycara,et al.  Transparency and Explanation in Deep Reinforcement Learning Neural Networks , 2018, AIES.

[23]  Tim Miller,et al.  Explanation in Artificial Intelligence: Insights from the Social Sciences , 2017, Artif. Intell..

[24]  Howard C. Warren Mental association from Plato to Hume. , 1916 .

[25]  David V. Pynadath,et al.  Building Trust in a Human-Robot Team with Automatically Generated Explanations , 2015 .

[26]  Ning Wang,et al.  Trust calibration within a human-robot team: Comparing automatically generated explanations , 2016, 2016 11th ACM/IEEE International Conference on Human-Robot Interaction (HRI).

[27]  Tim Miller,et al.  Explainable Reinforcement Learning Through a Causal Lens , 2019, AAAI.

[28]  Alessandra Sciutti,et al.  Learning from Learners: Adapting Reinforcement Learning Agents to be Competitive in a Card Game , 2020, 2020 25th International Conference on Pattern Recognition (ICPR).

[29]  Raymond Sheh,et al.  "Why Did You Do That?" Explainable Intelligent Robots , 2017, AAAI Workshops.

[30]  Abhinav Verma,et al.  Programmatically Interpretable Reinforcement Learning , 2018, ICML.

[31]  David Garlan,et al.  Toward Explainable Multi-Objective Probabilistic Planning , 2018, 2018 IEEE/ACM 4th International Workshop on Software Engineering for Smart Cyber-Physical Systems (SEsCPS).

[32]  Andrew Anderson,et al.  Explaining Reinforcement Learning to Mere Mortals: An Empirical Study , 2019, IJCAI.

[33]  Nikhil Churamani,et al.  iCub: Learning Emotion Expressions using Human Reward , 2020, ArXiv.

[34]  Liz Sonenberg,et al.  Distal Explanations for Explainable Reinforcement Learning Agents , 2020, ArXiv.

[35]  Katia P. Sycara,et al.  Object-sensitive Deep Reinforcement Learning , 2017, GCAI.

[36]  Xing Xie,et al.  A Reinforcement Learning Framework for Explainable Recommendation , 2018, 2018 IEEE International Conference on Data Mining (ICDM).

[37]  N. Daw,et al.  Reinforcement Learning and Episodic Memory in Humans and Animals: An Integrative Framework , 2017, Annual review of psychology.

[38]  Liz Sonenberg,et al.  Distal Explanations for Model-free Explainable Reinforcement Learning , 2020 .

[39]  Jessie Y. C. Chen,et al.  The influence of modality and transparency on trust in human-robot interaction , 2014, 2014 IEEE International Inter-Disciplinary Conference on Cognitive Methods in Situation Awareness and Decision Support (CogSIMA).

[40]  A. Cangelosi,et al.  Developmental Robotics: From Babies to Robots , 2015 .

[41]  Thomas A. Runkler,et al.  Interpretable Policies for Reinforcement Learning by Genetic Programming , 2017, Eng. Appl. Artif. Intell..

[42]  Bradley Hayes,et al.  Improving Human-Robot Interaction Through Explainable Reinforcement Learning , 2019, 2019 14th ACM/IEEE International Conference on Human-Robot Interaction (HRI).

[43]  Surya P. N. Singh,et al.  V-REP: A versatile and scalable robot simulation framework , 2013, 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[44]  Danilo Bassi,et al.  Indirect Training of Grey-Box Models: Application to a Bioprocess , 2007, ISNN.

[45]  Pedro Sequeira,et al.  Interestingness Elements for Explainable Reinforcement Learning: Understanding Agents' Capabilities and Limitations , 2019, Artif. Intell..

[46]  H. Pfister,et al.  How people explain their own and others’ behavior: a theory of lay causal explanations , 2015, Front. Psychol..

[47]  I. Pavlov Conditioned Reflexes: An Investigation of the Physiological Activity of the Cerebral Cortex , 1929 .

[48]  Eric M. S. P. Veith,et al.  Explainable Reinforcement Learning: A Survey , 2020, CD-MAKE.

[49]  Peter Vamplew,et al.  Memory-Based Explainable Reinforcement Learning , 2019, Australasian Conference on Artificial Intelligence.

[50]  Wojciech Jaskowski,et al.  ViZDoom: A Doom-based AI research platform for visual reinforcement learning , 2016, 2016 IEEE Conference on Computational Intelligence and Games (CIG).

[51]  Stefan Wermter,et al.  Explainable Goal-driven Agents and Robots - A Comprehensive Review , 2020, ACM Comput. Surv..

[52]  Gabriel Dulac-Arnold,et al.  Challenges of Real-World Reinforcement Learning , 2019, ArXiv.

[53]  Trevor Darrell,et al.  Generating Visual Explanations , 2016, ECCV.

[54]  Pat Langley,et al.  Explainable Agency for Intelligent Autonomous Systems , 2017, AAAI.