Recurrent World Models Facilitate Policy Evolution
暂无分享,去创建一个
[1] Richard S. Sutton,et al. Integrated Architectures for Learning, Planning, and Reacting Based on Approximating Dynamic Programming , 1990, ML.
[2] Jürgen Schmidhuber,et al. Learning to Forget: Continual Prediction with LSTM , 2000, Neural Computation.
[3] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[4] Pierre-Yves Oudeyer,et al. Information-seeking, curiosity, and attention: computational and neural mechanisms , 2013, Trends in Cognitive Sciences.
[5] Jürgen Schmidhuber,et al. Evolving large-scale neural networks for vision-based reinforcement learning , 2013, GECCO '13.
[6] Razvan Pascanu,et al. Visual Interaction Networks: Learning a Physics Simulator from Video , 2017, NIPS.
[7] Jürgen Schmidhuber,et al. Learning to Generate Artificial Fovea Trajectories for Target Detection , 1991, Int. J. Neural Syst..
[8] S. Hochreiter,et al. REINFORCEMENT DRIVEN INFORMATION ACQUISITION IN NONDETERMINISTIC ENVIRONMENTS , 1995 .
[9] Jürgen Schmidhuber,et al. First Experiments with PowerPlay , 2012, Neural networks : the official journal of the International Neural Network Society.
[10] Risto Miikkulainen,et al. Accelerated Neural Evolution through Cooperatively Coevolved Synapses , 2008, J. Mach. Learn. Res..
[11] Razvan Pascanu,et al. Imagination-Augmented Agents for Deep Reinforcement Learning , 2017, NIPS.
[12] Risto Miikkulainen,et al. Evolving Neural Networks through Augmenting Topologies , 2002, Evolutionary Computation.
[13] Razvan Pascanu,et al. Overcoming catastrophic forgetting in neural networks , 2016, Proceedings of the National Academy of Sciences.
[14] Jürgen Schmidhuber,et al. Co-evolving recurrent neurons learn deep memory POMDPs , 2005, GECCO '05.
[15] Thomas A. Runkler,et al. A benchmark environment motivated by industrial control problems , 2017, 2017 IEEE Symposium Series on Computational Intelligence (SSCI).
[16] Jürgen Schmidhuber,et al. Developmental robotics, optimal artificial curiosity, creativity, music, and the fine arts , 2006, Connect. Sci..
[17] Paul J. Werbos,et al. Neural networks for control and system identification , 1989, Proceedings of the 28th IEEE Conference on Decision and Control,.
[18] Joshua I. Sanders,et al. Cortical interneurons that specialize in disinhibitory control , 2013, Nature.
[19] Jürgen Schmidhuber,et al. An on-line algorithm for dynamic reinforcement learning and planning in reactive environments , 1990, 1990 IJCNN International Joint Conference on Neural Networks.
[20] Nikolaus Hansen,et al. The CMA Evolution Strategy: A Tutorial , 2016, ArXiv.
[21] Nikolaus Hansen,et al. Completely Derandomized Self-Adaptation in Evolution Strategies , 2001, Evolutionary Computation.
[22] Frank Fallside,et al. Dynamic reinforcement driven error propagation networks with application to game playing , 1989 .
[23] Alex Graves,et al. Playing Atari with Deep Reinforcement Learning , 2013, ArXiv.
[24] Jürgen Schmidhuber,et al. Learning Complex, Extended Sequences Using the Principle of History Compression , 1992, Neural Computation.
[25] Andrew W. Moore,et al. Reinforcement Learning: A Survey , 1996, J. Artif. Intell. Res..
[26] Wojciech Jaskowski,et al. ViZDoom: A Doom-based AI research platform for visual reinforcement learning , 2016, 2016 IEEE Conference on Computational Intelligence and Games (CIG).
[27] Henry J. Kelley,et al. Gradient Theory of Optimal Flight Paths , 1960 .
[28] Wojciech Zaremba,et al. OpenAI Gym , 2016, ArXiv.
[29] Quoc V. Le,et al. HyperNetworks , 2016, ICLR.
[30] Xi Chen,et al. Evolution Strategies as a Scalable Alternative to Reinforcement Learning , 2017, ArXiv.
[31] Thomas B. Schön,et al. From Pixels to Torques: Policy Learning with Deep Dynamical Models , 2015, ICML 2015.
[32] Carl E. Rasmussen,et al. Data-Efficient Reinforcement Learning in Continuous State-Action Gaussian-POMDPs , 2017, NIPS.
[33] Christopher Burgess,et al. DARLA: Improving Zero-Shot Transfer in Reinforcement Learning , 2017, ICML.
[34] Jürgen Schmidhuber,et al. Formal Theory of Creativity, Fun, and Intrinsic Motivation (1990–2010) , 2010, IEEE Transactions on Autonomous Mental Development.
[35] Eduardo F. Morales,et al. An Introduction to Reinforcement Learning , 2011 .
[36] Heiga Zen,et al. WaveNet: A Generative Model for Raw Audio , 2016, SSW.
[37] Sergey Levine,et al. Neural Network Dynamics for Model-Based Deep Reinforcement Learning with Model-Free Fine-Tuning , 2017, 2018 IEEE International Conference on Robotics and Automation (ICRA).
[38] Joseph Suarez,et al. Language Modeling with Recurrent Highway Hypernetworks , 2017, NIPS.
[39] Jürgen Schmidhuber,et al. On Learning to Think: Algorithmic Information Theory for Novel Combinations of Reinforcement Learning Controllers and Recurrent Neural World Models , 2015, ArXiv.
[40] Paul J. Werbos,et al. Applications of advances in nonlinear sensitivity analysis , 1982 .
[41] Vladlen Koltun,et al. Learning to Act by Predicting the Future , 2016, ICLR.
[42] Bobby D. Bryant,et al. Neurovisual Control in the Quake II Environment , 2012, IEEE Transactions on Computational Intelligence and AI in Games.
[43] Doris Y. Tsao,et al. The Code for Facial Identity in the Primate Brain , 2017, Cell.
[44] Sergey Levine,et al. Deep spatial autoencoders for visuomotor learning , 2015, 2016 IEEE International Conference on Robotics and Automation (ICRA).
[45] Robert M. French,et al. Catastrophic Interference in Connectionist Networks: Can It Be Predicted, Can It Be Prevented? , 1993, NIPS.
[46] Frank Sehnke,et al. Parameter-exploring policy gradients , 2010, Neural Networks.
[47] Vighnesh Birodkar,et al. Unsupervised Learning of Disentangled Representations from Video , 2017, NIPS.
[48] Jürgen Schmidhuber,et al. PowerPlay: Training an Increasingly General Problem Solver by Continually Searching for the Simplest Still Unsolvable Problem , 2011, Front. Psychol..
[49] Alex Graves,et al. Generating Sequences With Recurrent Neural Networks , 2013, ArXiv.
[50] Daan Wierstra,et al. Stochastic Backpropagation and Approximate Inference in Deep Generative Models , 2014, ICML.
[51] Jürgen Schmidhuber,et al. Curious model-building control systems , 1991, [Proceedings] 1991 IEEE International Joint Conference on Neural Networks.
[52] P. König,et al. Primary Visual Cortex Represents the Difference Between Past and Present , 2013, Cerebral cortex.
[53] Boyang Li,et al. Game Engine Learning from Video , 2017, IJCAI.
[54] Honglak Lee,et al. Action-Conditional Video Prediction using Deep Networks in Atari Games , 2015, NIPS.
[55] Ian Johnson,et al. Experiments in Handwriting with a Neural Network , 2016 .
[56] David Whitney,et al. Motion-Dependent Representation of Space in Area MT+ , 2013, Neuron.
[57] Kenneth O. Stanley,et al. Abandoning Objectives: Evolution Through the Search for Novelty Alone , 2011, Evolutionary Computation.
[58] Max Welling,et al. Auto-Encoding Variational Bayes , 2013, ICLR.
[59] C. Koch,et al. Invariant visual representation by single neurons in the human brain , 2005, Nature.
[60] Kenneth O. Stanley,et al. Autonomous Evolution of Topographic Regularities in Artificial Neural Networks , 2010, Neural Computation.
[61] R Ratcliff,et al. Connectionist models of recognition memory: constraints imposed by learning and forgetting functions. , 1990, Psychological review.
[62] C. Rasmussen,et al. Improving PILCO with Bayesian Neural Network Dynamics Models , 2016 .
[63] Geoffrey E. Hinton,et al. Outrageously Large Neural Networks: The Sparsely-Gated Mixture-of-Experts Layer , 2017, ICLR.
[64] T. Sejnowski,et al. Nanoconnectomic upper bound on the variability of synaptic plasticity , 2015, eLife.
[65] Risto Miikkulainen,et al. A Neuroevolution Approach to General Atari Game Playing , 2014, IEEE Transactions on Computational Intelligence and AI in Games.
[66] B. Widrow,et al. The truck backer-upper: an example of self-learning in neural networks , 1989, International 1989 Joint Conference on Neural Networks.
[67] Tom Schaul,et al. The Predictron: End-To-End Learning and Planning , 2016, ICML.
[68] Alexei A. Efros,et al. Curiosity-Driven Exploration by Self-Supervised Prediction , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).
[69] Finale Doshi-Velez,et al. Learning and Policy Search in Stochastic Dynamical Systems with Bayesian Neural Networks , 2016, ICLR.
[70] Pierre-Yves Oudeyer,et al. Intrinsic Motivation Systems for Autonomous Mental Development , 2007, IEEE Transactions on Evolutionary Computation.
[71] Sergey Levine,et al. Using Simulation and Domain Adaptation to Improve Efficiency of Deep Robotic Grasping , 2018, 2018 IEEE International Conference on Robotics and Automation (ICRA).
[72] Alex Graves,et al. The Kanerva Machine: A Generative Distributed Memory , 2018, ICLR.
[73] Douglas Eck,et al. A Neural Representation of Sketch Drawings , 2017, ICLR.
[74] Georg B. Keller,et al. A Sensorimotor Circuit in Mouse Cortex for Visual Flow Predictions , 2017, Neuron.
[75] Jürgen Schmidhuber,et al. One Big Net For Everything , 2018, ArXiv.
[76] David Amos,et al. Generative Temporal Models with Memory , 2017, ArXiv.
[77] Long-Ji Lin,et al. Reinforcement learning for robots using neural networks , 1992 .
[78] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[79] Kenneth O. Stanley,et al. Deep Neuroevolution: Genetic Algorithms Are a Competitive Alternative for Training Deep Neural Networks for Reinforcement Learning , 2017, ArXiv.
[80] Georg B. Keller,et al. Sensorimotor Mismatch Signals in Primary Visual Cortex of the Behaving Mouse , 2012, Neuron.
[81] Thomas B. Schön,et al. Learning deep dynamical models from image pixels , 2014, ArXiv.
[82] Chrisantha Fernando,et al. PathNet: Evolution Channels Gradient Descent in Super Neural Networks , 2017, ArXiv.
[83] Julian Togelius,et al. Autoencoder-augmented neuroevolution for visual doom playing , 2017, 2017 IEEE Conference on Computational Intelligence and Games (CIG).
[84] Daan Wierstra,et al. Recurrent Environment Simulators , 2017, ICLR.
[85] Sophia Blau,et al. Numerical Optimization Of Computer Models , 2016 .
[86] Carl E. Rasmussen,et al. PILCO: A Model-Based and Data-Efficient Approach to Policy Search , 2011, ICML.
[87] Martin A. Riedmiller,et al. Embed to Control: A Locally Linear Latent Dynamics Model for Control from Raw Images , 2015, NIPS.
[88] Jürgen Schmidhuber,et al. A possibility for implementing curiosity and boredom in model-building neural controllers , 1991 .
[89] D. Mobbs,et al. The ecology of human fear: survival optimization and the nervous system , 2015, Front. Neurosci..
[90] Jürgen Schmidhuber,et al. Reinforcement Learning in Markovian and Non-Markovian Environments , 1990, NIPS.
[91] David Ha,et al. Evolving Stable Strategies , 2017 .