Explainable Goal-driven Agents and Robots - A Comprehensive Review

Recent applications of autonomous agents and robots have brought attention to crucial trust-related challenges associated with the current generation of artificial intelligence (AI) systems. AI systems based on the connectionist deep learning neural network approach lack capabilities of explaining their decisions and actions to others, despite their great successes. Without symbolic interpretation capabilities, they are ‘black boxes’, which renders their choices or actions opaque, making it difficult to trust them in safety-critical applications. The recent stance on the explainability of AI systems has witnessed several approaches to eXplainable Artificial Intelligence (XAI); however, most of the studies have focused on data-driven XAI systems applied in computational sciences. Studies addressing the increasingly pervasive goal-driven agents and robots are sparse at this point in time. This paper reviews approaches on explainable goal-driven intelligent agents and robots, focusing on techniques for explaining and communicating agents’ perceptual functions (e.g., senses, vision) and cognitive reasoning (e.g., beliefs, desires, intentions, plans, and goals) with humans in the loop. The review highlights key strategies that emphasize transparency, understandability, and continual learning for explainability. Finally, the paper presents requirements for explainability and suggests a road map for the possible realization of effective goal-driven explainable agents and robots.

[1]  S. Wermter,et al.  What’s on Your Mind, NICO? , 2022, KI - Künstliche Intelligenz.

[2]  S. Wermter,et al.  Explain yourself! Effects of Explanations in Human-Robot Interaction , 2022, 2022 31st IEEE International Conference on Robot and Human Interactive Communication (RO-MAN).

[3]  Amanda Prorok,et al.  Explanation-Aware Experience Replay in Rule-Dense Environments , 2021, IEEE Robotics and Automation Letters.

[4]  Stefan Wermter,et al.  Emotion Recognition Using Explainable Genetically Optimized Fuzzy ART Ensembles , 2021, IEEE Access.

[5]  Stefan Wermter,et al.  Tell Me Why You Feel That Way: Processing Compositional Dependency for Tree-LSTM Aspect Sentiment Triplet Extraction (TASTE) , 2021, ICANN.

[6]  Nina Dethlefs,et al.  XAI4Wind: A Multimodal Knowledge Graph Database for Explainable Decision Support in Operations & Maintenance of Wind Turbines , 2020, ArXiv.

[7]  Chien-Min Wang,et al.  Explainable and Adaptable Augmentation in Knowledge Attention Network for Multi-Agent Deep Reinforcement Learning Systems , 2020, 2020 IEEE Third International Conference on Artificial Intelligence and Knowledge Engineering (AIKE).

[8]  Lionel M. Ni,et al.  HeTROPY: Explainable learning diagnostics via heterogeneous maximum-entropy and multi-spatial knowledge representation , 2020, Knowl. Based Syst..

[9]  Francisco Cruz,et al.  Unmanned Aerial Vehicle Control Through Domain-based Automatic Speech Recognition , 2020, Comput..

[10]  Joe Townsend,et al.  Extracting Relational Explanations From Deep Neural Networks: A Survey From a Neural-Symbolic Perspective , 2020, IEEE Transactions on Neural Networks and Learning Systems.

[11]  Paul Davidsson,et al.  An Agent-Based Approach to Realize Emergent Configurations in the Internet of Things , 2020, Electronics.

[12]  A. Kori,et al.  Abstracting Deep Neural Networks into Concept Graphs for Concept Level Interpretability , 2020, 2008.06457.

[13]  Xiaofeng Gao,et al.  Joint Mind Modeling for Explanation Generation in Complex Human-Robot Collaborative Tasks , 2020, 2020 29th IEEE International Conference on Robot and Human Interactive Communication (RO-MAN).

[14]  Sercan O. Arik,et al.  Explaining Deep Neural Networks using Unsupervised Clustering , 2020, ArXiv.

[15]  Arun Das,et al.  Opportunities and Challenges in Explainable Artificial Intelligence (XAI): A Survey , 2020, ArXiv.

[16]  Alberto Del Bimbo,et al.  Explaining Autonomous Driving by Learning End-to-End Visual Attention , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).

[17]  Chang Wang,et al.  Coactive design of explainable agent-based task planning and deep reinforcement learning for human-UAVs teamwork , 2020 .

[18]  Mohit Bansal,et al.  Evaluating Explainable AI: Which Algorithmic Explanations Help Users Predict Model Behavior? , 2020, ACL.

[19]  Paolo Fiorini,et al.  Autonomous task planning and situation awareness in robotic surgery* , 2020, 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[20]  Zheng-Yu Niu,et al.  Knowledge Graph Grounded Goal Planning for Open-Domain Conversation Generation , 2020, AAAI.

[21]  Nykan Mirchi,et al.  The Virtual Operative Assistant: An explainable artificial intelligence tool for simulation-based training in surgery and medicine , 2020, PloS one.

[22]  Enrico Costanza,et al.  Evaluating saliency map explanations for convolutional neural networks: a user study , 2020, IUI.

[23]  Sven Magg,et al.  Efficient Facial Feature Learning with Wide Ensemble-based Convolutional Neural Networks , 2020, AAAI.

[24]  Mohit Shridhar,et al.  INGRESS: Interactive visual grounding of referring expressions , 2020, Int. J. Robotics Res..

[25]  M. Gervasio,et al.  Interestingness Elements for Explainable Reinforcement Learning: Understanding Agents' Capabilities and Limitations , 2019, Artif. Intell..

[26]  Yujun Zeng,et al.  Deep, Consistent Behavioral Decision Making with Planning Features for Autonomous Vehicles , 2019, Electronics.

[27]  Bo Thiesson,et al.  Explainable artificial intelligence model to predict acute critical illness from electronic health records , 2019, Nature Communications.

[28]  Peter Vamplew,et al.  Memory-Based Explainable Reinforcement Learning , 2019, Australasian Conference on Artificial Intelligence.

[29]  Stefan Wermter,et al.  Bootstrapping Knowledge Graphs From Images and Text , 2019, Front. Neurorobot..

[30]  Kim Veltman,et al.  Intelligent autonomous vehicles with an extendable knowledge base under meaningful human control , 2019, Security + Defence.

[31]  Sehoon Ha,et al.  Zero-shot Imitation Learning from Demonstrations for Legged Robot Visual Navigation , 2019, 2020 IEEE International Conference on Robotics and Automation (ICRA).

[32]  Ben Meadows,et al.  Towards a Theory of Explanations for Human–Robot Collaboration , 2019, KI - Künstliche Intelligenz.

[33]  Iván García-Magariño,et al.  Human-Centric AI for Trustworthy IoT Systems With Explainable Multilayer Perceptrons , 2019, IEEE Access.

[34]  Alan Fern,et al.  Strategic Tasks for Explainable Reinforcement Learning , 2019, AAAI.

[35]  Tim Miller,et al.  Explainable Reinforcement Learning Through a Causal Lens , 2019, AAAI.

[36]  Mathieu Bourgais,et al.  BEN: An Agent Architecture for Explainable and Expressive Behavior in Social Simulation , 2019, EXTRAAMAS@AAMAS.

[37]  Davide Calvaresi,et al.  Explainable Agents and Robots: Results from a Systematic Literature Review , 2019, AAMAS.

[38]  Luciano Serafini,et al.  Neural-Symbolic Computing: An Effective Methodology for Principled Integration of Machine Learning and Reasoning , 2019, FLAP.

[39]  Melanie Mitchell,et al.  Revisiting Visual Grounding , 2019, Proceedings of the Second Workshop on Shortcomings in Vision and Language.

[40]  Levent Yilmaz,et al.  A Cognitive Architecture for Verifiable System Ethics via Explainable Autonomy , 2019, 2019 IEEE International Systems Conference (SysCon).

[41]  Bradley Hayes,et al.  Improving Human-Robot Interaction Through Explainable Reinforcement Learning , 2019, 2019 14th ACM/IEEE International Conference on Human-Robot Interaction (HRI).

[42]  Shalini Ghosh,et al.  Generating Natural Language Explanations for Visual Question Answering using Scene Graphs and Visual Attention , 2019, ArXiv.

[43]  Mark O. Riedl,et al.  Automated rationale generation: a technique for explainable AI and its effects on human perceptions , 2019, IUI.

[44]  Tao Li,et al.  Visual Interrogation of Attention-Based Models for Natural Language Inference and Machine Comprehension , 2018, EMNLP.

[45]  Stefan Wermter,et al.  Classification of MRI Migraine Medical Data Using 3D Convolutional Neural Network , 2018, ICANN.

[46]  Helen F. Hastie,et al.  MIRIAM: A Multimodal Interface for Explaining the Reasoning Behind Actions of Remote Autonomous Systems , 2018, ICMI.

[47]  Alun D. Preece,et al.  Stakeholders in Explainable AI , 2018, ArXiv.

[48]  Matthew Guzdial,et al.  Explainable PCGML via Game Design Patterns , 2018, AIIDE Workshops.

[49]  Mica R. Endsley,et al.  Innovative model for situation awareness in dynamic defense systems , 2018 .

[50]  Yu Zhang,et al.  Behavior Explanation as Intention Signaling in Human-Robot Teaming , 2018, 2018 27th IEEE International Symposium on Robot and Human Interactive Communication (RO-MAN).

[51]  Stefan Lee,et al.  Graph R-CNN for Scene Graph Generation , 2018, ECCV.

[52]  Shanshan Zhang,et al.  Interpretable Representation Learning for Healthcare via Capturing Disease Progression through Time , 2018, KDD.

[53]  Mark A. Neerincx,et al.  Using Perceptual and Cognitive Explanations for Enhanced Human-Agent Team Performance , 2018, HCI.

[54]  Daniele Magazzeni,et al.  Towards Providing Explanations for AI Planner Decisions , 2018, IJCAI 2018.

[55]  Samir Aknine,et al.  How explainable plans can make planning faster , 2018 .

[56]  Ning Wang,et al.  Clustering Behavior to Recognize Subjective Beliefs in Human-Agent Teams , 2018, AAMAS.

[57]  Ofra Amir,et al.  HIGHLIGHTS: Summarizing Agent Behavior to People , 2018, AAMAS.

[58]  Amit Dhurandhar,et al.  Teaching Meaningful Explanations , 2018, ArXiv.

[59]  David Garlan,et al.  Toward Explainable Multi-Objective Probabilistic Planning , 2018, 2018 IEEE/ACM 4th International Workshop on Software Engineering for Smart Cyber-Physical Systems (SEsCPS).

[60]  Luowei Zhou,et al.  Weakly-Supervised Video Object Grounding from Text by Loss Weighting and Object Interaction , 2018, BMVC.

[61]  Talal Rahwan,et al.  How AI Wins Friends and Influences People in Repeated Games With Cheap Talk , 2018, AAAI.

[62]  Seiji Yamada,et al.  Effect of Expressive Lights on Human Perception and Interpretation of Functional Robot , 2018, CHI Extended Abstracts.

[63]  Theresa-Marie Rhyne,et al.  Visual Analytics for Explainable Deep Learning , 2018, IEEE Computer Graphics and Applications.

[64]  Masahiko Mikawa,et al.  Expression of intention by rotational head movements for teleoperated mobile robot , 2018, 2018 IEEE 15th International Workshop on Advanced Motion Control (AMC).

[65]  Jessie Y. C. Chen,et al.  Situation awareness-based agent transparency and human-autonomy teaming effectiveness , 2018 .

[66]  Franco Turini,et al.  A Survey of Methods for Explaining Black Box Models , 2018, ACM Comput. Surv..

[67]  Andreas Holzinger,et al.  Towards the Augmented Pathologist: Challenges of Explainable-AI in Digital Pathology , 2017, ArXiv.

[68]  Koen V. Hindriks,et al.  Personalised self-explanation by robots: The role of goals versus beliefs in robot-action explanation for children and adults , 2017, 2017 26th IEEE International Symposium on Robot and Human Interactive Communication (RO-MAN).

[69]  Eran Yahav,et al.  Extracting Automata from Recurrent Neural Networks Using Queries and Counterexamples , 2017, ICML.

[70]  Koen V. Hindriks,et al.  The role of emotion in self-explanations by cognitive agents , 2017, 2017 Seventh International Conference on Affective Computing and Intelligent Interaction Workshops and Demos (ACIIW).

[71]  Jessie Y. C. Chen,et al.  Effects of agent transparency and communication framing on human-agent teaming , 2017, 2017 IEEE International Conference on Systems, Man, and Cybernetics (SMC).

[72]  Rachel K. E. Bellamy,et al.  Visualizations for an Explainable Planning Agent , 2017, IJCAI.

[73]  Jun-Seok Kim,et al.  Interactive Visualization and Manipulation of Attention-based Neural Machine Translation , 2017, EMNLP.

[74]  Joanna Bryson,et al.  Improving robot transparency: Real-time visualisation of robot AI substantially improves understanding in naive observers , 2017, 2017 26th IEEE International Symposium on Robot and Human Interactive Communication (RO-MAN).

[75]  Pieter Abbeel,et al.  Learning Generalized Reactive Policies using Deep Neural Networks , 2017, ICAPS.

[76]  May D. Wang,et al.  Interpretable Predictions of Clinical Outcomes with An Attention-based Recurrent Neural Network , 2017, BCB.

[77]  Juan A. Recio-García,et al.  Make it personal: A social explanation system applied to group recommendations , 2017, Expert Syst. Appl..

[78]  Luciano Floridi,et al.  Transparent, explainable, and accountable AI for robotics , 2017, Science Robotics.

[79]  B. Scassellati,et al.  Social eye gaze in human-robot interaction , 2017, J. Hum. Robot Interact..

[80]  Scott Lundberg,et al.  A Unified Approach to Interpreting Model Predictions , 2017, NIPS.

[81]  Rhona Flin,et al.  Incident Command: Tales from the Hot Seat , 2017 .

[82]  Anca D. Dragan,et al.  Translating Neuralese , 2017, ACL.

[83]  Devi Parikh,et al.  It Takes Two to Tango: Towards Theory of AI's Mind , 2017, ArXiv.

[84]  Bradley Hayes,et al.  Improving Robot Controller Transparency Through Autonomous Policy Explanation , 2017, 2017 12th ACM/IEEE International Conference on Human-Robot Interaction (HRI.

[85]  Mark O. Riedl,et al.  Rationalization: A Neural Machine Translation Approach to Generating Natural Language Explanations , 2017, AIES.

[86]  Pat Langley,et al.  Explainable Agency for Intelligent Autonomous Systems , 2017, AAAI.

[87]  Yu Zhang,et al.  Plan Explanations as Model Reconciliation: Moving Beyond Explanation as Soliloquy , 2017, IJCAI.

[88]  Danfei Xu,et al.  Scene Graph Generation by Iterative Message Passing , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[89]  Licheng Yu,et al.  A Joint Speaker-Listener-Reinforcer Model for Referring Expressions , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[90]  David W. Aha,et al.  Incorporating Transparency During Trust-Guided Behavior Adaptation , 2016, ICCBR.

[91]  Alan Bundy,et al.  Preparing for the future of Artificial Intelligence , 2016, AI & SOCIETY.

[92]  Ramprasaath R. Selvaraju,et al.  Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization , 2016, International Journal of Computer Vision.

[93]  Dietmar Jannach,et al.  Recommendation quality, transparency, and website quality for trust-building in recommendation agents , 2016, Electron. Commer. Res. Appl..

[94]  Jimeng Sun,et al.  RETAIN: An Interpretable Predictive Model for Healthcare using Reverse Time Attention Mechanism , 2016, NIPS.

[95]  Stephanie Rosenthal,et al.  Verbalization: Narration of Autonomous Robot Experience , 2016, IJCAI.

[96]  Andreas Theodorou,et al.  What Does the Robot Think? Transparency as a Fundamental Design Requirement for Intelligent Systems , 2016, IJCAI 2016.

[97]  Marco Mirolli,et al.  GRAIL: A Goal-Discovering Robotic Architecture for Intrinsically-Motivated Learning , 2016, IEEE Transactions on Cognitive and Developmental Systems.

[98]  Ning Wang,et al.  The Impact of POMDP-Generated Explanations on Trust and Performance in Human-Robot Teams , 2016, AAMAS.

[99]  Julian N. Marewski,et al.  What can the brain teach us about building artificial intelligence? , 2016, Behavioral and Brain Sciences.

[100]  Volker Tresp,et al.  Predicting Clinical Events by Combining Static and Dynamic Information Using Recurrent Neural Networks , 2016, 2016 IEEE International Conference on Healthcare Informatics (ICHI).

[101]  Bolei Zhou,et al.  Learning Deep Features for Discriminative Localization , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[102]  Charles Elkan,et al.  Learning to Diagnose with LSTM Recurrent Neural Networks , 2015, ICLR.

[103]  Elizabeth Sklar,et al.  Learning Spatial Models for Navigation , 2015, COSIT.

[104]  Ravi Teja Chadalavada,et al.  That's on my mind! robot to human intention communication through on-board projection on shared floor space , 2015, 2015 European Conference on Mobile Robots (ECMR).

[105]  Jekaterina Novikova,et al.  Emotionally expressive robot behavior improves human-robot collaboration , 2015, 2015 24th IEEE International Symposium on Robot and Human Interactive Communication (RO-MAN).

[106]  Michael S. Bernstein,et al.  Image retrieval using scene graphs , 2015, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[107]  Kaiming He,et al.  Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks , 2015, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[108]  Tim Miller,et al.  Communication in Human-Agent Teams for Tasks with Joint Action , 2015, COIN@AAMAS/IJCAI.

[109]  Siddhartha S. Srinivasa,et al.  Effects of Robot Motion on Human-Robot Collaboration , 2015, 2015 10th ACM/IEEE International Conference on Human-Robot Interaction (HRI).

[110]  Jessie Y. C. Chen,et al.  Effects of Agent Transparency on Operator Trust , 2015, HRI.

[111]  V. S. Costa,et al.  Inductive Logic Programming , 2014, Lecture Notes in Computer Science.

[112]  Bernd Schattenberg,et al.  From Abstract Crisis to Concrete Relief — A Preliminary Report on Combining State Abstraction and HTN Planning , 2014 .

[113]  Michael W. Boyce,et al.  Situation Awareness-Based Agent Transparency , 2014 .

[114]  Weng-Keen Wong,et al.  Too much, too little, or just right? Ways explanations impact end users' mental models , 2013, 2013 IEEE Symposium on Visual Languages and Human Centric Computing.

[115]  Philippe Gaussier,et al.  From self-assessment to frustration, a small step toward autonomy in robotic navigation , 2013, Front. Neurorobot..

[116]  Changsong Liu,et al.  Towards Situated Dialogue: Revisiting Referring Expression Generation , 2013, EMNLP.

[117]  Axel Schulte,et al.  Self-explanation capability for cognitive agents on-board of UCAVs to improve cooperation in a manned-unmanned fighter team , 2013 .

[118]  Bilge Mutlu,et al.  Theory of mind: mechanisms, methods, and new directions , 2013, Front. Hum. Neurosci..

[119]  Khaled Ghedira,et al.  Explanation language syntax for Multi-Agent Systems , 2013, 2013 World Congress on Computer and Information Technology (WCCIT).

[120]  D. Aha,et al.  GOAL‐DRIVEN AUTONOMY FOR RESPONDING TO UNEXPECTED EVENTS IN STRATEGY SIMULATIONS , 2013, Comput. Intell..

[121]  Simone Stumpf,et al.  The effect of explanations on perceived control and behaviors in intelligent systems , 2013, CHI Extended Abstracts.

[122]  Ana Paiva,et al.  Social Robots for Long-Term Interaction: A Survey , 2013, International Journal of Social Robotics.

[123]  Arnav Jhala,et al.  Learning from Demonstration for Goal-Driven Autonomy , 2012, AAAI.

[124]  Ielka van der Sluis,et al.  Generation of Referring Expressions: Assessing the Incremental Algorithm , 2012, Cogn. Sci..

[125]  E. Vincent Cross,et al.  Explaining robot actions , 2012, 2012 7th ACM/IEEE International Conference on Human-Robot Interaction (HRI).

[126]  Michael D. Coovert,et al.  Exploration of intention expression for robots , 2012, 2012 7th ACM/IEEE International Conference on Human-Robot Interaction (HRI).

[127]  Emiel Krahmer,et al.  Computational Generation of Referring Expressions: A Survey , 2012, CL.

[128]  Frank Puppe,et al.  Systematic introduction to expert systems - knowledge representations and problem-solving methods , 2011 .

[129]  Anind K. Dey,et al.  Design of an intelligible mobile context-aware application , 2011, Mobile HCI.

[130]  Jorge A. Baier,et al.  Preferred Explanations: Theory and Generation via Planning , 2011, AAAI.

[131]  Takayuki Kanda,et al.  A conversational robot in an elderly care center: An ethnographic study , 2011, 2011 6th ACM/IEEE International Conference on Human-Robot Interaction (HRI).

[132]  Wendy Ju,et al.  Expressing thought: Improving robot readability with animation principles , 2011, 2011 6th ACM/IEEE International Conference on Human-Robot Interaction (HRI).

[133]  Koen V. Hindriks,et al.  Do You Get It? User-Evaluated Explainable BDI Agents , 2010, MATES.

[134]  Jo Vermeulen,et al.  Improving intelligibility and control in Ubicomp , 2010, UbiComp '10 Adjunct.

[135]  Weng-Keen Wong,et al.  Explanatory Debugging: Supporting End-User Debugging of Machine-Learned Programs , 2010, VL/HCC.

[136]  John-Jules Ch. Meyer,et al.  Design and Evaluation of Explainable BDI Agents , 2010, 2010 IEEE/WIC/ACM International Conference on Web Intelligence and Intelligent Agent Technology.

[137]  Mary Ellen Foster,et al.  Situated Reference in a Hybrid Human-Robot Interaction System , 2010, INLG.

[138]  David W. Aha,et al.  Goal-Driven Autonomy in a Navy Strategy Simulation , 2010, AAAI.

[139]  Takeo Kanade,et al.  The Extended Cohn-Kanade Dataset (CK+): A complete dataset for action unit and emotion-specified expression , 2010, 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition - Workshops.

[140]  John-Jules Ch. Meyer,et al.  A Methodology for Developing Self-Explaining Agents for Virtual Training , 2009, MALLOW.

[141]  Allen Munro,et al.  Towards Research on Goal Reasoning with the TAO Sandbox , 2009 .

[142]  Yang Wang,et al.  Robot Navigation by Waypoints , 2008, J. Intell. Robotic Syst..

[143]  B. Wandell What's in your mind? , 2008, Nature Neuroscience.

[144]  Alois Knoll,et al.  The roles of haptic-ostensive referring expressions in cooperative, task-based human-robot dialogue , 2008, 2008 3rd ACM/IEEE International Conference on Human-Robot Interaction (HRI).

[145]  Stefan Wermter,et al.  Auto-Extraction, Representation and Integration of a Diabetes Ontology Using Bayesian Networks , 2007, Twentieth IEEE International Symposium on Computer-Based Medical Systems (CBMS'07).

[146]  Michael T. Cox Perpetual Self-Aware Cognitive Agents , 2007, AI Mag..

[147]  Pamela J. Hinds,et al.  Who Should I Blame? Effects of Autonomy and Transparency on Attributions in Human-Robot Interaction , 2006, ROMAN 2006 - The 15th IEEE International Symposium on Robot and Human Interactive Communication.

[148]  Miquel Sànchez-Marrè,et al.  A purely reactive navigation scheme for dynamic environments using Case-Based Reasoning , 2006, Auton. Robots.

[149]  Stefan Wermter,et al.  Data mining using rule extraction from Kohonen self-organising maps , 2006, Neural Computing & Applications.

[150]  A. Sugiyama,et al.  Auditory system in a personal robot, PaPeRo , 2006, 2006 Digest of Technical Papers International Conference on Consumer Electronics.

[151]  Andrea Lockerd Thomaz,et al.  Effects of nonverbal communication on efficiency and robustness in human-robot teamwork , 2005, 2005 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[152]  Frank Stajano,et al.  Visually Interactive Location-Aware Computing , 2005, UbiComp.

[153]  N. Block Two neural correlates of consciousness , 2005, Trends in Cognitive Sciences.

[154]  Michael van Lent,et al.  An Explainable Artificial Intelligence System for Small-unit Tactical Behavior , 2004, AAAI.

[155]  Peter Carey,et al.  Data Protection: A Practical Guide to UK and EU Law , 2004 .

[156]  D. Barr,et al.  Limits on theory of mind use in adults , 2003, Cognition.

[157]  Sabine A. Döring Explaining Action by Emotion , 2003 .

[158]  Stefan Wermter,et al.  Symbolic state transducers and recurrent neural preference machines for text mining , 2003, Int. J. Approx. Reason..

[159]  Jean Scholtz,et al.  Theory and evaluation of human robot interactions , 2003, 36th Annual Hawaii International Conference on System Sciences, 2003. Proceedings of the.

[160]  Kiyohiro Shikano,et al.  ASKA: receptionist robot with speech dialogue system , 2002, IEEE/RSJ International Conference on Intelligent Robots and Systems.

[161]  Kikuo Fujimura,et al.  The intelligent ASIMO: system overview and integration , 2002, IEEE/RSJ International Conference on Intelligent Robots and Systems.

[162]  Stefan Wermter,et al.  Knowledge Extraction from Local Function Networks , 2001, IJCAI.

[163]  Ashwin Ram,et al.  Introspective Multistrategy Learning: On the Construction of Learning Strategies , 1999, Artif. Intell..

[164]  Hector Muñoz-Avila,et al.  SHOP: Simple Hierarchical Ordered Planner , 1999, IJCAI.

[165]  Stefan Wermter,et al.  Preference Moore Machines for Neural Fuzzy Integration , 1999, IJCAI.

[166]  John Bigham,et al.  Software Agents for Future Communication Systems , 1999, Springer Berlin Heidelberg.

[167]  B. Malle,et al.  How People Explain Behavior: A New Theoretical Framework , 1999, Personality and social psychology review : an official journal of the Society for Personality and Social Psychology, Inc.

[168]  Catriona Kennedy,et al.  A conceptual foundation for autonomous learning in unforeseen situations , 1998, Proceedings of the 1998 IEEE International Symposium on Intelligent Control (ISIC) held jointly with IEEE International Symposium on Computational Intelligence in Robotics and Automation (CIRA) Intell.

[169]  Michael Wooldridge,et al.  The Belief-Desire-Intention Model of Agency , 1998, ATAL.

[170]  Joachim Diederich,et al.  Survey and critique of techniques for extracting rules from trained artificial neural networks , 1995, Knowl. Based Syst..

[171]  Robert Dale,et al.  Computational Interpretations of the Gricean Maxims in the Generation of Referring Expressions , 1995, Cogn. Sci..

[172]  Sonia Fernandez Preparing for the future , 1995, Veterinary Record.

[173]  Michael K. Sahota Reactive Deliberation: An Architecture for Real-Time Intelligent Control in Dynamic Environments , 1994, AAAI.

[174]  W. Lewis Johnson,et al.  Agents that Learn to Explain Themselves , 1994, AAAI.

[175]  Michael L. Littman,et al.  Memoryless policies: theoretical limitations and practical results , 1994 .

[176]  Subbarao Kambhampati,et al.  A Unified Framework for Explanation-Based Generalization of Partially Ordered and Partially Instantiated Plans , 1994, Artif. Intell..

[177]  Padhraic Smyth,et al.  Learning Finite State Machines With Self-Clustering Recurrent Networks , 1993, Neural Computation.

[178]  Kevin Knight,et al.  Are Many Reactive Agents Better Than a Few Deliberative Ones? , 1993, IJCAI.

[179]  Johanna D. Moore,et al.  Explanation in second generation expert systems , 1993 .

[180]  Long Ji Lin,et al.  Self-improving reactive agents based on reinforcement learning, planning and teaching , 1992, Machine Learning.

[181]  C. Lee Giles,et al.  Learning and Extracting Finite State Automata with Second-Order Recurrent Neural Networks , 1992, Neural Computation.

[182]  Michael E. Bratman,et al.  Intention, Plans, and Practical Reason , 1991 .

[183]  T. Kohonen The self-organizing map , 1990, Neurocomputing.

[184]  Johan de Kleer,et al.  An Assumption-Based TMS , 1987, Artif. Intell..

[185]  Scott Bennett,et al.  A Domain Independent Explanation-Based Generalizer , 1986, AAAI.

[186]  Katsuhiko Shirai,et al.  SPEECH CONVERSATION SYSTEM OF THE MUSICIAN ROBOT. , 1985 .

[187]  Katsuhiko Shirai,et al.  Ear and voice of the WABOT The speech-input-output system (ワボット特集) , 1973 .

[188]  Richard Fikes,et al.  STRIPS: A New Approach to the Application of Theorem Proving to Problem Solving , 1971, IJCAI.

[189]  Lorenzo Feruglio,et al.  On applying AI-driven flight data analysis for operational spacecraft model-based diagnostics , 2020, Annu. Rev. Control..

[190]  Dongil Han,et al.  A Deep Learning-Based Hybrid Framework for Object Detection and Recognition in Autonomous Driving , 2020, IEEE Access.

[191]  Mauro Da Lio,et al.  Agent Architecture for Adaptive Behaviors in Autonomous Driving , 2020, IEEE Access.

[192]  Alan Fern,et al.  Explainable Reinforcement Learning via Reward Decomposition , 2019 .

[193]  Sepp Hochreiter,et al.  Visual Scene Understanding for Autonomous Driving Using Semantic Segmentation , 2019, Explainable AI.

[194]  Eric Yeh,et al.  Explanation to Avert Surprise , 2018, IUI Workshops.

[195]  David W. Aha,et al.  Towards Explainable NPCs: A Relational Exploration Learning Agent , 2018, AAAI Workshops.

[196]  Michael W. Floyd,et al.  Learning from Exploration: Towards an Explainable Goal Reasoning Agent , 2018 .

[197]  M. J. Robeer,et al.  Contrastive Explanation for Machine Learning , 2018 .

[198]  Mark A. Neerincx,et al.  ICM: An Intuitive Model Independent and Accurate Certainty Measure for Machine Learning , 2018, ICAART.

[199]  H. Hermanns,et al.  From Machine Ethics To Machine Explainability and Back , 2018, ISAIM.

[200]  Claude Sammut,et al.  Towards Explainable Tool Creation by a Robot , 2017 .

[201]  Or Biran,et al.  Explanation and Justification in Machine Learning : A Survey Or , 2017 .

[202]  Ana Paiva,et al.  Expressive Lights for Revealing Mobile Service Robot State , 2015, ROBOT.

[203]  Khaled Ghédira,et al.  International Conference in Knowledge Based and Intelligent Information and Engineering Systems-KES 2013 Intra-agent explanation using temporal and extended causal maps , 2013 .

[204]  Willem F. G. Haselager,et al.  Signaling Robot Trustworthiness: Effects of Behavioral Cues as Warnings , 2013, ICSR 2013.

[205]  Weng-Keen Wong,et al.  Making intelligent systems understandable and controllable by end users , 2012 .

[206]  D. Aha,et al.  Case-Based Learning in Goal-Driven Autonomy Agents for Real-Time Strategy Combat Tasks , 2011 .

[207]  Wolfgang Minker,et al.  Verbal Plan Explanations for Hybrid Planning , 2010, MKWI.

[208]  John-Jules Ch. Meyer,et al.  Guidelines for developing explainable cognitive models , 2010 .

[209]  David W. Aha,et al.  A Case Study of Goal-Driven Autonomy in Domination Games , 2010 .

[210]  Frank E. Ritter,et al.  Designs for explaining intelligent agents , 2009, Int. J. Hum. Comput. Stud..

[211]  Cynthia Breazeal,et al.  Social Robots that Interact with People , 2008, Springer Handbook of Robotics.

[212]  Dana S. Nau Automated Planning , 2008, Handbook of Knowledge Representation.

[213]  Shinichi Ohnaka,et al.  Scenario and design process of childcare robot PaPeRo , 2006, ACE '06.

[214]  F. Keil,et al.  Explanation and understanding , 2015 .

[215]  Ipke Wachsmuth,et al.  Incremental Generation of Multimodal Deixis Referring to Objects , 2005, ENLG.

[216]  Stefan Wermter,et al.  Knowledge Extraction from Transducer Neural Networks , 2004, Applied Intelligence.

[217]  Robert Dale,et al.  Handbook of Natural Language Processing , 2001, Computational Linguistics.

[218]  R. Stainton Perspectives in the Philosophy of Language: A Concise Anthology , 2000 .

[219]  Thomas Rist,et al.  Rocco: A RoboCup Soccer Commentator System , 1998, RoboCup.

[220]  Paul Davidsson,et al.  Autonomous Agents and the Concept of Concepts , 1996 .

[221]  C. Lee Giles,et al.  Extraction of rules from discrete-time recurrent neural networks , 1996, Neural Networks.

[222]  Michael Wooldridge,et al.  Conceptualising and Developing Agents , 1995 .

[223]  M. Georgeff,et al.  BDI Agents: From Theory to Practice , 1995, ICMAS.

[224]  Logic Programming: PROLOG. , 1989 .

[225]  Janet L. Kolodner,et al.  Case-Based Reasoning , 1988, IJCAI 1989.

[226]  I. Dekieer An assumption-based truth maintenance system , 1986 .

[227]  Proceedings of the Twenty-Second International Joint Conference on Artificial Intelligence Integrated Learning for Goal-Driven Autonomy , 2022 .

[228]  Proceedings of the Twenty-Second International Conference on Automated Planning and Scheduling Making Hybrid Plans More Clear to Human Users — A Formal Approach for Generating Sound Explanations , 2022 .