Tell me why! Explanations support learning relational and causal structure

Inferring the abstract relational and causal structure of the world is a major challenge for reinforcement-learning (RL) agents. For humans, language—particularly in the form of explanations —plays a considerable role in overcoming this challenge. Here, we show that language can play a similar role for deep RL agents in complex environments. While agents typically struggle to acquire relational and causal knowledge, augmenting their experience by training them to predict language descriptions and explanations can overcome these limitations. We show that language can help agents learn challenging relational tasks, and examine which aspects of language contribute to its benefits. We then show that explanations can help agents to infer not only relational but also causal structure. Language can shape the way that agents to generalize out-of-distribution from ambiguous, causally-confounded training, and explanations even allow agents to learn to perform experimental interventions to identify causal relationships. Our results suggest that language description and explanation may be powerful tools for improving agent learning and generalization. It is often argued that machine learning models—and deep learning models in particular—lack the human proficiencies forming abstractions and inferring relational or causal These limitations can make it hard to train models that generalize out-of-distribution, or that reason in human-like ways, particularly for reinforcement learning (RL) agents that receive high-bandwidth input from raw pixels and must learn to act in partially-observable environments. the main text results without explanations show that these losses are not sufficient for learning either. This shows that the benefits of explanations are not simply due to having more supervision for the agent, but rather are specific to supervision that highlights the abstract task structure.

[1]  Andrew Kyle Lampinen,et al.  Semantic Exploration from Language Abstractions and Pretrained Representations , 2022, ArXiv.

[2]  Improving Intrinsic Exploration with Language Abstractions , 2022, ArXiv.

[3]  Mohit Bansal,et al.  When Can Models Learn From Explanations? A Formal Framework for Understanding the Roles of Explanation Data , 2021, LNLS.

[4]  Marcel van Gerven,et al.  Explainable Deep Learning: A Field Guide for the Uninitiated , 2020, J. Artif. Intell. Res..

[5]  P. Abbeel,et al.  Teachable Reinforcement Learning via Advice Distillation , 2022, NeurIPS.

[6]  A Survey of Generalisation in Deep Reinforcement Learning , 2021, ArXiv.

[7]  Oriol Vinyals,et al.  Highly accurate protein structure prediction with AlphaFold , 2021, Nature.

[8]  Sergey Levine,et al.  Why Generalization in RL is Difficult: Epistemic POMDPs and Implicit Partial Observability , 2021, NeurIPS.

[9]  James L. McClelland,et al.  What underlies rapid learning and systematic generalization in humans , 2021, ArXiv.

[10]  Alan Yuille,et al.  Visual analogy: Deep learning versus compositional models , 2021, ArXiv.

[11]  Francesca Toni,et al.  Explanation-Based Human Debugging of NLP Models: A Survey , 2021, Transactions of the Association for Computational Linguistics.

[12]  J. Bowers,et al.  Can Deep Convolutional Neural Networks Learn Same-Different Relations? , 2021, bioRxiv.

[13]  Andrew Kyle Lampinen,et al.  Symbolic Behaviour in Artificial Intelligence , 2021, ArXiv.

[14]  K. Holyoak,et al.  Emergence of relational reasoning , 2021, Current Opinion in Behavioral Sciences.

[15]  A. Wright,et al.  Issues in the comparative cognition of same/different abstract-concept learning , 2021, Current Opinion in Behavioral Sciences.

[16]  S. Gershman,et al.  Memory as a Computational Resource , 2021, Trends in Cognitive Sciences.

[17]  K. Kersting,et al.  Right for the Right Concept: Revising Neuro-Symbolic Concepts by Interacting with their Explanations , 2020, 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[18]  Mudit Verma,et al.  Widening the Pipeline in Human-Guided Reinforcement Learning with Explanation and Context-Aware Data Augmentation , 2020, NeurIPS.

[19]  Zeb Kurth-Nelson,et al.  Alchemy: A structured task distribution for meta-reinforcement learning , 2021, ArXiv.

[20]  Luis C. Lamb,et al.  Neurosymbolic AI: the 3rd wave , 2020, Artificial Intelligence Review.

[21]  N. Thompson Metaphysical Explanation , 2020, The Routledge Handbook of Metametaphysics.

[22]  Ilya Kostrikov,et al.  Automatic Data Augmentation for Generalization in Deep Reinforcement Learning , 2020, ArXiv.

[23]  Andrew Kyle Lampinen,et al.  What shapes feature representations? Exploring datasets, architectures, and training , 2020, NeurIPS.

[24]  Christopher Potts,et al.  Relational reasoning and generalization using non-symbolic neural networks , 2020, CogSci.

[25]  M. Bethge,et al.  Shortcut learning in deep neural networks , 2020, Nature Machine Intelligence.

[26]  Gary Marcus,et al.  The Next Decade in AI: Four Steps Towards Robust Artificial Intelligence , 2020, ArXiv.

[27]  Fabio Viola,et al.  Causally Correct Partial Models for Reinforcement Learning , 2020, ArXiv.

[28]  Kristian Kersting,et al.  Making deep neural networks right for the right scientific reasons by interacting with their explanations , 2020, Nat. Mach. Intell..

[29]  Noah D. Goodman,et al.  Shaping Visual Representations with Language for Few-Shot Classification , 2019, ACL.

[30]  Razvan Pascanu,et al.  Stabilizing Transformers for Reinforcement Learning , 2019, ICML.

[31]  Oleg O. Sushkov,et al.  Scaling data-driven robotics with reward sketching and batch reinforcement learning , 2019, Robotics: Science and Systems.

[32]  James L. McClelland,et al.  Environmental drivers of systematicity and generalization in a situated agent , 2019, ICLR.

[33]  Murray Shanahan,et al.  An Explicitly Relational Neural Network Architecture , 2019, ICML.

[34]  Francisco S. Melo,et al.  Learning from Explanations and Demonstrations: A Pilot Study , 2020, NL4XAI.

[35]  Guy Dove More than a scaffold: Language is a neuroenhancement , 2020, Cognitive neuropsychology.

[36]  Aaron van den Oord,et al.  Shaping Belief States with Generative Environment Models for RL , 2019, NeurIPS.

[37]  Chelsea Finn,et al.  Language as an Abstraction for Hierarchical Deep Reinforcement Learning , 2019, NeurIPS.

[38]  Shimon Whiteson,et al.  A Survey of Reinforcement Learning Informed by Natural Language , 2019, IJCAI.

[39]  Manuela Veloso,et al.  Generation of Policy-Level Explanations for Reinforcement Learning , 2019, AAAI.

[40]  Yee Whye Teh,et al.  Meta reinforcement learning as task inference , 2019, ArXiv.

[41]  D. Gentner,et al.  Explanation recruits comparison in a category-learning task , 2019, Cognition.

[42]  Sergey Levine,et al.  Efficient Off-Policy Meta-Reinforcement Learning via Probabilistic Context Variables , 2019, ICML.

[43]  Prasoon Goyal,et al.  Using Natural Language for Reward Shaping in Reinforcement Learning , 2019, IJCAI.

[44]  Judea Pearl,et al.  The seven tools of causal inference, with reflections on machine learning , 2019, Commun. ACM.

[45]  Zeb Kurth-Nelson,et al.  Causal Reasoning from Meta-reinforcement Learning , 2019, ArXiv.

[46]  Thomas Lukasiewicz,et al.  e-SNLI: Natural Language Inference with Natural Language Explanations , 2018, NeurIPS.

[47]  Felix Hill,et al.  Measuring abstract reasoning in neural networks , 2018, ICML.

[48]  Le Song,et al.  Learning to Explain: An Information-Theoretic Perspective on Model Interpretation , 2018, ICML.

[49]  Shane Legg,et al.  IMPALA: Scalable Distributed Deep-RL with Importance Weighted Actor-Learner Architectures , 2018, ICML.

[50]  Judea Pearl,et al.  Theoretical Impediments to Machine Learning With Seven Sparks from the Causal Revolution , 2018, WSDM.

[51]  Dan Klein,et al.  Learning with Latent Language , 2017, NAACL.

[52]  Demis Hassabis,et al.  Grounded Language Learning in a Simulated 3D World , 2017, ArXiv.

[53]  Razvan Pascanu,et al.  A simple neural network module for relational reasoning , 2017, NIPS.

[54]  Chris Sauer,et al.  Beating Atari with Natural Language Guided Reinforcement Learning , 2017, ArXiv.

[55]  Andrew Slavin Ross,et al.  Right for the Right Reasons: Training Differentiable Models by Constraining their Explanations , 2017, IJCAI.

[56]  Tom Schaul,et al.  Reinforcement Learning with Unsupervised Auxiliary Tasks , 2016, ICLR.

[57]  Michael R. Waldmann,et al.  The Oxford handbook of causal reasoning , 2017 .

[58]  G. Lupyan The Centrality of Language in Human Cognition , 2016 .

[59]  Joshua B. Tenenbaum,et al.  Building machines that learn and think like people , 2016, Behavioral and Brain Sciences.

[60]  G. Lupyan,et al.  What makes words special? Words as unmotivated cues , 2015, Cognition.

[61]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[62]  Robert L. Goldstone,et al.  Concreteness Fading in Mathematics and Science Instruction: a Systematic Review , 2014 .

[63]  Jivko Sinapov,et al.  The odd one out task: Toward an intelligence test for robots , 2010, 2010 IEEE 9th International Conference on Development and Learning.

[64]  Joseph Jay Williams,et al.  The role of explanation in discovery and generalization: evidence from category learning , 2010, ICLS.

[65]  E. Warrington,et al.  The different representational frameworks underpinning abstract and concrete knowledge: Evidence from odd-one-out judgements , 2009, Quarterly journal of experimental psychology.

[66]  Dedre Gentner,et al.  Relational language supports relational cognition in humans and apes , 2008, Behavioral and Brain Sciences.

[67]  Derek C. Penn,et al.  Darwin's mistake: Explaining the discontinuity between human and nonhuman minds , 2008, Behavioral and Brain Sciences.

[68]  G. Lupyan Taking symbols for granted? Is the discontinuity between human and nonhuman minds the product of external symbol systems? , 2008, Behavioral and Brain Sciences.

[69]  Daniel J. Navarro,et al.  One of these greebles is not like the others: Semi-supervised models for similarity structures , 2008 .

[70]  Keith A. Markus,et al.  Making Things Happen: A Theory of Causal Explanation , 2007 .

[71]  T. Lombrozo The structure and function of explanations , 2006, Trends in Cognitive Sciences.

[72]  S. Carey,et al.  Functional explanation and the function of explanation , 2006, Cognition.

[73]  B. Rittle-Johnson,et al.  Promoting transfer: effects of self-explanation and direct instruction. , 2006, Child development.

[74]  Dedre Gentner,et al.  Why we’re so smart , 2003 .

[75]  D. Gentner,et al.  Language in Mind: Advances in the Study of Language and Thought , 2003 .

[76]  David M. Sobel,et al.  Detecting blickets: how young children use information about novel causal powers in categorization and induction. , 2000, Child development.

[77]  Robert A. Wilson,et al.  Explanation and Cognition , 2000 .

[78]  A. Gopnik,et al.  The scientist in the crib : minds, brains, and how children learn , 1999 .

[79]  Michelene T. H. Chi,et al.  Eliciting Self-Explanations Improves Understanding , 1994, Cogn. Sci..

[80]  R. Mooney,et al.  Schema acquisition from a single example , 1992 .

[81]  J. Fodor,et al.  Connectionism and cognitive architecture: A critical analysis , 1988, Cognition.

[82]  Raimo Tuomela,et al.  A Pragmatic Theory of Explanation , 1984 .

[83]  J. Bruner,et al.  The role of tutoring in problem solving. , 1976, Journal of child psychology and psychiatry, and allied disciplines.