Analogues of mental simulation and imagination in deep learning
暂无分享,去创建一个
[1] Amir Dezfouli,et al. Speed/Accuracy Trade-Off between the Habitual and the Goal-Directed Processes , 2011, PLoS Comput. Biol..
[2] Wei Ji Ma,et al. Do People Think Like Computers? , 2016, Computers and Games.
[3] Richard S. Sutton,et al. Integrated Architectures for Learning, Planning, and Reacting Based on Approximating Dynamic Programming , 1990, ML.
[4] Jitendra Malik,et al. Learning to Poke by Poking: Experiential Learning of Intuitive Physics , 2016, NIPS.
[5] Wulfram Gerstner,et al. Efficient Model-Based Deep Reinforcement Learning with Variational State Tabulation , 2018, ICML.
[6] Rob Fergus,et al. Learning Physical Intuition of Block Towers by Example , 2016, ICML.
[7] Joel Z. Leibo,et al. Prefrontal cortex as a meta-reinforcement learning system , 2018, bioRxiv.
[8] Sean R Eddy,et al. What is dynamic programming? , 2004, Nature Biotechnology.
[9] Razvan Pascanu,et al. Relational inductive biases, deep learning, and graph networks , 2018, ArXiv.
[10] Anders Sandberg,et al. Blueberry Earth , 2018, Physics World.
[11] R. Shepard,et al. Mental Rotation of Three-Dimensional Objects , 1971, Science.
[12] K. Doya,et al. A unifying computational framework for motor control and social interaction. , 2003, Philosophical transactions of the Royal Society of London. Series B, Biological sciences.
[13] Z. Pylyshyn. Mental imagery: In search of a theory , 2002, Behavioral and Brain Sciences.
[14] R. N. Spreng,et al. The Future of Memory: Remembering, Imagining, and the Brain , 2012, Neuron.
[15] Rémi Munos,et al. Learning to Search with MCTSnets , 2018, ICML.
[16] Bernt Schiele,et al. Long-Term Image Boundary Prediction , 2016, AAAI.
[17] M. Kunda. Visual mental imagery: A view from artificial intelligence , 2018, Cortex.
[18] R A Finke,et al. Explorations of creative visual synthesis in mental imagery , 1988, Memory & cognition.
[19] Richard S. Sutton,et al. Reinforcement Learning , 1992, Handbook of Machine Learning.
[20] Jennifer Curtis,et al. What Happens if ... , 2011 .
[21] Fabio Viola,et al. Learning and Querying Fast Generative Models for Reinforcement Learning , 2018, ArXiv.
[22] Jessica B. Hamrick,et al. Simulation as an engine of physical scene understanding , 2013, Proceedings of the National Academy of Sciences.
[23] D. Hassabis,et al. Neuroscience-Inspired Artificial Intelligence , 2017, Neuron.
[24] Razvan Pascanu,et al. Learning model-based planning from scratch , 2017, ArXiv.
[25] Tom Schaul,et al. The Predictron: End-To-End Learning and Planning , 2016, ICML.
[26] Jitendra Malik,et al. Learning Visual Predictive Models of Physics for Playing Billiards , 2015, ICLR.
[27] Daan Wierstra,et al. Recurrent Environment Simulators , 2017, ICLR.
[28] J. Dauben. Georg Cantor and the Origins of Transfinite Set Theory , 1983 .
[29] P. N. Johnson-Laird,et al. Inference with Mental Models , 2012 .
[30] W. H. F. Barnes. The Nature of Explanation , 1944, Nature.
[31] Daniel L. K. Yamins,et al. Flexible Neural Representation for Physics Prediction , 2018, NeurIPS.
[32] Rick Grush,et al. The emulation theory of representation: Motor control, imagery, and perception , 2004, Behavioral and Brain Sciences.
[33] Joshua B. Tenenbaum,et al. Building machines that learn and think like people , 2016, Behavioral and Brain Sciences.
[34] Konrad P. Körding,et al. Toward an Integration of Deep Learning and Neuroscience , 2016, bioRxiv.
[35] Pieter Abbeel,et al. Learning Plannable Representations with Causal InfoGAN , 2018, NeurIPS.
[36] Satinder Singh,et al. Value Prediction Network , 2017, NIPS.
[37] S. Kosslyn,et al. Mental imagery , 2013, Front. Psychol..
[38] Jürgen Schmidhuber,et al. Relational Neural Expectation Maximization: Unsupervised Discovery of Objects and their Interactions , 2018, ICLR.
[39] D. Hassabis,et al. Hippocampal place cells construct reward related sequences through unexplored space , 2015, eLife.
[40] Tania Lombrozo,et al. “Learning by Thinking” in Science and in Everyday Life , 2020 .
[41] Allan Jabri,et al. Universal Planning Networks , 2018, ICML.
[42] Kevin J. Miller,et al. Dorsal hippocampus contributes to model-based planning , 2017, Nature Neuroscience.
[43] Sergey Levine,et al. Continuous Deep Q-Learning with Model-based Acceleration , 2016, ICML.
[44] M. Hegarty. Mechanical reasoning by mental simulation , 2004, Trends in Cognitive Sciences.
[45] Demis Hassabis,et al. Mastering the game of Go with deep neural networks and tree search , 2016, Nature.
[46] Yuval Tassa,et al. Learning Continuous Control Policies by Stochastic Value Gradients , 2015, NIPS.
[47] Sergey Levine,et al. Deep visual foresight for planning robot motion , 2016, 2017 IEEE International Conference on Robotics and Automation (ICRA).
[48] Marcelo G Mattar,et al. Prioritized memory access explains planning and hippocampal replay , 2017, Nature Neuroscience.
[49] Bernard W. Balleine,et al. Actions, Action Sequences and Habits: Evidence That Goal-Directed and Habitual Action Control Are Hierarchically Organized , 2013, PLoS Comput. Biol..
[50] Leslie Pack Kaelbling,et al. Planning and Acting in Partially Observable Stochastic Domains , 1998, Artif. Intell..
[51] S. Kosslyn,et al. Mental imagery , 2013, Front. Psychol..
[52] Ali Farhadi,et al. "What Happens If..." Learning to Predict the Effect of Forces in Images , 2016, ECCV.
[53] J. Tenenbaum,et al. Intuitive Theories , 2020, Encyclopedia of Creativity, Invention, Innovation and Entrepreneurship.
[54] Jiajun Wu,et al. Learning to See Physics via Visual De-animation , 2017, NIPS.
[55] J. DiCarlo,et al. Using goal-driven deep learning models to understand sensory cortex , 2016, Nature Neuroscience.
[56] K. Holyoak,et al. The Oxford handbook of thinking and reasoning , 2012 .
[57] Jürgen Schmidhuber,et al. Recurrent World Models Facilitate Policy Evolution , 2018, NeurIPS.
[58] Raia Hadsell,et al. Graph networks as learnable physics engines for inference and control , 2018, ICML.
[59] Rolf A. Zwaan. Situation Models , 1999 .
[60] Jiajun Wu,et al. Galileo: Perceiving Physical Object Properties by Integrating a Physics Engine with Deep Learning , 2015, NIPS.
[61] Yedid Hoshen,et al. VAIN: Attentional Multi-agent Predictive Modeling , 2017, NIPS.
[62] Joshua B. Tenenbaum,et al. A Compositional Object-Based Approach to Learning Physical Dynamics , 2016, ICLR.
[63] D. Hassabis,et al. Using Imagination to Understand the Neural Basis of Episodic Memory , 2007, The Journal of Neuroscience.
[64] David Hsu,et al. QMDP-Net: Deep Learning for Planning under Partial Observability , 2017, NIPS.
[65] Shimon Whiteson,et al. Deep Variational Reinforcement Learning for POMDPs , 2018, ICML.
[66] Marc Peter Deisenroth,et al. Deep Reinforcement Learning: A Brief Survey , 2017, IEEE Signal Processing Magazine.
[67] Rob Fergus,et al. Composable Planning with Attributes , 2018, ICML.
[68] M. L. Story. Learning by thinking , 1953 .
[69] Sergey Levine,et al. Neural Network Dynamics for Model-Based Deep Reinforcement Learning with Model-Free Fine-Tuning , 2017, 2018 IEEE International Conference on Robotics and Automation (ICRA).
[70] Thomas L. Griffiths,et al. Evaluating (and Improving) the Correspondence Between Deep Neural Networks and Human Representations , 2017, Cogn. Sci..
[71] Razvan Pascanu,et al. Visual Interaction Networks: Learning a Physics Simulator from Video , 2017, NIPS.
[72] Demis Hassabis,et al. Mastering the game of Go without human knowledge , 2017, Nature.
[73] P. Harris. The work of imagination , 1991 .
[74] Razvan Pascanu,et al. Metacontrol for Adaptive Imagination-Based Optimization , 2017, ICLR.
[75] Sergey Levine,et al. Deep Reinforcement Learning in a Handful of Trials using Probabilistic Dynamics Models , 2018, NeurIPS.
[76] Carolyn Copper,et al. Does mental practice enhance performance , 1994 .
[77] Shimon Whiteson,et al. TreeQN and ATreeC: Differentiable Tree Planning for Deep Reinforcement Learning , 2017, ICLR 2018.
[78] Razvan Pascanu,et al. Imagination-Augmented Agents for Deep Reinforcement Learning , 2017, NIPS.
[79] P. Dayan,et al. Model-based influences on humans’ choices and striatal prediction errors , 2011, Neuron.
[80] Samuel Gershman,et al. Imaginative Reinforcement Learning: Computational Principles and Neural Mechanisms , 2017, Journal of Cognitive Neuroscience.
[81] Pieter Abbeel,et al. Value Iteration Networks , 2016, NIPS.
[82] Robert J. Zatorre,et al. Mental Concerts: Musical Imagery and Auditory Cortex , 2005, Neuron.
[83] Daniel L. K. Yamins,et al. Common Object Representations for Visual Production and Recognition , 2018, Cogn. Sci..
[84] Razvan Pascanu,et al. Interaction Networks for Learning about Objects, Relations and Physics , 2016, NIPS.
[85] P. Dayan,et al. Adaptive integration of habits into depth-limited planning defines a habitual-goal–directed spectrum , 2016, Proceedings of the National Academy of Sciences.
[86] Anne Kuefer,et al. The Case For Mental Imagery , 2016 .
[87] M. Jeannerod. Mental imagery in the motor context , 1995, Neuropsychologia.
[88] Michael I. Jordan,et al. Model-Based Value Estimation for Efficient Model-Free Reinforcement Learning , 2018, ArXiv.
[89] John J. Clement,et al. The Role of Imagistic Simulation in Scientific Thought Experiments , 2009 .
[90] Chris L. Baker,et al. Modeling Human Plan Recognition Using Bayesian Theory of Mind , 2014 .