Analogues of mental simulation and imagination in deep learning

Mental simulation — the capacity to imagine what will or what could be — is a salient feature of human cognition, playing a key role in a wide range of cognitive abilities. In artificial intelligence, the last few years have seen the development of methods which are analogous to mental models and mental simulation. This paper outlines recent methods in deep learning for constructing such models from data and learning to use them via reinforcement learning, and compares such approaches to human mental simulation. Model-based methods in deep learning can serve as powerful tools for building and scaling cognitive models. However, a number of challenges remain in matching the capacity of human mental simulation for efficiency, compositionality, generalization, and creativity.

[1]  Amir Dezfouli,et al.  Speed/Accuracy Trade-Off between the Habitual and the Goal-Directed Processes , 2011, PLoS Comput. Biol..

[2]  Wei Ji Ma,et al.  Do People Think Like Computers? , 2016, Computers and Games.

[3]  Richard S. Sutton,et al.  Integrated Architectures for Learning, Planning, and Reacting Based on Approximating Dynamic Programming , 1990, ML.

[4]  Jitendra Malik,et al.  Learning to Poke by Poking: Experiential Learning of Intuitive Physics , 2016, NIPS.

[5]  Wulfram Gerstner,et al.  Efficient Model-Based Deep Reinforcement Learning with Variational State Tabulation , 2018, ICML.

[6]  Rob Fergus,et al.  Learning Physical Intuition of Block Towers by Example , 2016, ICML.

[7]  Joel Z. Leibo,et al.  Prefrontal cortex as a meta-reinforcement learning system , 2018, bioRxiv.

[8]  Sean R Eddy,et al.  What is dynamic programming? , 2004, Nature Biotechnology.

[9]  Razvan Pascanu,et al.  Relational inductive biases, deep learning, and graph networks , 2018, ArXiv.

[10]  Anders Sandberg,et al.  Blueberry Earth , 2018, Physics World.

[11]  R. Shepard,et al.  Mental Rotation of Three-Dimensional Objects , 1971, Science.

[12]  K. Doya,et al.  A unifying computational framework for motor control and social interaction. , 2003, Philosophical transactions of the Royal Society of London. Series B, Biological sciences.

[13]  Z. Pylyshyn Mental imagery: In search of a theory , 2002, Behavioral and Brain Sciences.

[14]  R. N. Spreng,et al.  The Future of Memory: Remembering, Imagining, and the Brain , 2012, Neuron.

[15]  Rémi Munos,et al.  Learning to Search with MCTSnets , 2018, ICML.

[16]  Bernt Schiele,et al.  Long-Term Image Boundary Prediction , 2016, AAAI.

[17]  M. Kunda Visual mental imagery: A view from artificial intelligence , 2018, Cortex.

[18]  R A Finke,et al.  Explorations of creative visual synthesis in mental imagery , 1988, Memory & cognition.

[19]  Richard S. Sutton,et al.  Reinforcement Learning , 1992, Handbook of Machine Learning.

[20]  Jennifer Curtis,et al.  What Happens if ... , 2011 .

[21]  Fabio Viola,et al.  Learning and Querying Fast Generative Models for Reinforcement Learning , 2018, ArXiv.

[22]  Jessica B. Hamrick,et al.  Simulation as an engine of physical scene understanding , 2013, Proceedings of the National Academy of Sciences.

[23]  D. Hassabis,et al.  Neuroscience-Inspired Artificial Intelligence , 2017, Neuron.

[24]  Razvan Pascanu,et al.  Learning model-based planning from scratch , 2017, ArXiv.

[25]  Tom Schaul,et al.  The Predictron: End-To-End Learning and Planning , 2016, ICML.

[26]  Jitendra Malik,et al.  Learning Visual Predictive Models of Physics for Playing Billiards , 2015, ICLR.

[27]  Daan Wierstra,et al.  Recurrent Environment Simulators , 2017, ICLR.

[28]  J. Dauben Georg Cantor and the Origins of Transfinite Set Theory , 1983 .

[29]  P. N. Johnson-Laird,et al.  Inference with Mental Models , 2012 .

[30]  W. H. F. Barnes The Nature of Explanation , 1944, Nature.

[31]  Daniel L. K. Yamins,et al.  Flexible Neural Representation for Physics Prediction , 2018, NeurIPS.

[32]  Rick Grush,et al.  The emulation theory of representation: Motor control, imagery, and perception , 2004, Behavioral and Brain Sciences.

[33]  Joshua B. Tenenbaum,et al.  Building machines that learn and think like people , 2016, Behavioral and Brain Sciences.

[34]  Konrad P. Körding,et al.  Toward an Integration of Deep Learning and Neuroscience , 2016, bioRxiv.

[35]  Pieter Abbeel,et al.  Learning Plannable Representations with Causal InfoGAN , 2018, NeurIPS.

[36]  Satinder Singh,et al.  Value Prediction Network , 2017, NIPS.

[37]  S. Kosslyn,et al.  Mental imagery , 2013, Front. Psychol..

[38]  Jürgen Schmidhuber,et al.  Relational Neural Expectation Maximization: Unsupervised Discovery of Objects and their Interactions , 2018, ICLR.

[39]  D. Hassabis,et al.  Hippocampal place cells construct reward related sequences through unexplored space , 2015, eLife.

[40]  Tania Lombrozo,et al.  “Learning by Thinking” in Science and in Everyday Life , 2020 .

[41]  Allan Jabri,et al.  Universal Planning Networks , 2018, ICML.

[42]  Kevin J. Miller,et al.  Dorsal hippocampus contributes to model-based planning , 2017, Nature Neuroscience.

[43]  Sergey Levine,et al.  Continuous Deep Q-Learning with Model-based Acceleration , 2016, ICML.

[44]  M. Hegarty Mechanical reasoning by mental simulation , 2004, Trends in Cognitive Sciences.

[45]  Demis Hassabis,et al.  Mastering the game of Go with deep neural networks and tree search , 2016, Nature.

[46]  Yuval Tassa,et al.  Learning Continuous Control Policies by Stochastic Value Gradients , 2015, NIPS.

[47]  Sergey Levine,et al.  Deep visual foresight for planning robot motion , 2016, 2017 IEEE International Conference on Robotics and Automation (ICRA).

[48]  Marcelo G Mattar,et al.  Prioritized memory access explains planning and hippocampal replay , 2017, Nature Neuroscience.

[49]  Bernard W. Balleine,et al.  Actions, Action Sequences and Habits: Evidence That Goal-Directed and Habitual Action Control Are Hierarchically Organized , 2013, PLoS Comput. Biol..

[50]  Leslie Pack Kaelbling,et al.  Planning and Acting in Partially Observable Stochastic Domains , 1998, Artif. Intell..

[51]  S. Kosslyn,et al.  Mental imagery , 2013, Front. Psychol..

[52]  Ali Farhadi,et al.  "What Happens If..." Learning to Predict the Effect of Forces in Images , 2016, ECCV.

[53]  J. Tenenbaum,et al.  Intuitive Theories , 2020, Encyclopedia of Creativity, Invention, Innovation and Entrepreneurship.

[54]  Jiajun Wu,et al.  Learning to See Physics via Visual De-animation , 2017, NIPS.

[55]  J. DiCarlo,et al.  Using goal-driven deep learning models to understand sensory cortex , 2016, Nature Neuroscience.

[56]  K. Holyoak,et al.  The Oxford handbook of thinking and reasoning , 2012 .

[57]  Jürgen Schmidhuber,et al.  Recurrent World Models Facilitate Policy Evolution , 2018, NeurIPS.

[58]  Raia Hadsell,et al.  Graph networks as learnable physics engines for inference and control , 2018, ICML.

[59]  Rolf A. Zwaan Situation Models , 1999 .

[60]  Jiajun Wu,et al.  Galileo: Perceiving Physical Object Properties by Integrating a Physics Engine with Deep Learning , 2015, NIPS.

[61]  Yedid Hoshen,et al.  VAIN: Attentional Multi-agent Predictive Modeling , 2017, NIPS.

[62]  Joshua B. Tenenbaum,et al.  A Compositional Object-Based Approach to Learning Physical Dynamics , 2016, ICLR.

[63]  D. Hassabis,et al.  Using Imagination to Understand the Neural Basis of Episodic Memory , 2007, The Journal of Neuroscience.

[64]  David Hsu,et al.  QMDP-Net: Deep Learning for Planning under Partial Observability , 2017, NIPS.

[65]  Shimon Whiteson,et al.  Deep Variational Reinforcement Learning for POMDPs , 2018, ICML.

[66]  Marc Peter Deisenroth,et al.  Deep Reinforcement Learning: A Brief Survey , 2017, IEEE Signal Processing Magazine.

[67]  Rob Fergus,et al.  Composable Planning with Attributes , 2018, ICML.

[68]  M. L. Story Learning by thinking , 1953 .

[69]  Sergey Levine,et al.  Neural Network Dynamics for Model-Based Deep Reinforcement Learning with Model-Free Fine-Tuning , 2017, 2018 IEEE International Conference on Robotics and Automation (ICRA).

[70]  Thomas L. Griffiths,et al.  Evaluating (and Improving) the Correspondence Between Deep Neural Networks and Human Representations , 2017, Cogn. Sci..

[71]  Razvan Pascanu,et al.  Visual Interaction Networks: Learning a Physics Simulator from Video , 2017, NIPS.

[72]  Demis Hassabis,et al.  Mastering the game of Go without human knowledge , 2017, Nature.

[73]  P. Harris The work of imagination , 1991 .

[74]  Razvan Pascanu,et al.  Metacontrol for Adaptive Imagination-Based Optimization , 2017, ICLR.

[75]  Sergey Levine,et al.  Deep Reinforcement Learning in a Handful of Trials using Probabilistic Dynamics Models , 2018, NeurIPS.

[76]  Carolyn Copper,et al.  Does mental practice enhance performance , 1994 .

[77]  Shimon Whiteson,et al.  TreeQN and ATreeC: Differentiable Tree Planning for Deep Reinforcement Learning , 2017, ICLR 2018.

[78]  Razvan Pascanu,et al.  Imagination-Augmented Agents for Deep Reinforcement Learning , 2017, NIPS.

[79]  P. Dayan,et al.  Model-based influences on humans’ choices and striatal prediction errors , 2011, Neuron.

[80]  Samuel Gershman,et al.  Imaginative Reinforcement Learning: Computational Principles and Neural Mechanisms , 2017, Journal of Cognitive Neuroscience.

[81]  Pieter Abbeel,et al.  Value Iteration Networks , 2016, NIPS.

[82]  Robert J. Zatorre,et al.  Mental Concerts: Musical Imagery and Auditory Cortex , 2005, Neuron.

[83]  Daniel L. K. Yamins,et al.  Common Object Representations for Visual Production and Recognition , 2018, Cogn. Sci..

[84]  Razvan Pascanu,et al.  Interaction Networks for Learning about Objects, Relations and Physics , 2016, NIPS.

[85]  P. Dayan,et al.  Adaptive integration of habits into depth-limited planning defines a habitual-goal–directed spectrum , 2016, Proceedings of the National Academy of Sciences.

[86]  Anne Kuefer,et al.  The Case For Mental Imagery , 2016 .

[87]  M. Jeannerod Mental imagery in the motor context , 1995, Neuropsychologia.

[88]  Michael I. Jordan,et al.  Model-Based Value Estimation for Efficient Model-Free Reinforcement Learning , 2018, ArXiv.

[89]  John J. Clement,et al.  The Role of Imagistic Simulation in Scientific Thought Experiments , 2009 .

[90]  Chris L. Baker,et al.  Modeling Human Plan Recognition Using Bayesian Theory of Mind , 2014 .