暂无分享,去创建一个
Amos J. Storkey | Harrison Edwards | Oleg Klimov | Yuri Burda | Oleg Klimov | A. Storkey | Yuri Burda | Harrison Edwards
[1] Stewart W. Wilson,et al. A Possibility for Implementing Curiosity and Boredom in Model-Building Neural Controllers , 1991 .
[2] Jürgen Schmidhuber,et al. Curious model-building control systems , 1991, [Proceedings] 1991 IEEE International Joint Conference on Neural Networks.
[3] Benjamin Recht,et al. Random Features for Large-Scale Kernel Machines , 2007, NIPS.
[4] Pierre-Yves Oudeyer,et al. Intrinsic Motivation Systems for Autonomous Mental Development , 2007, IEEE Transactions on Evolutionary Computation.
[5] Michael L. Littman,et al. An analysis of model-based Interval Estimation for Markov Decision Processes , 2008, J. Comput. Syst. Sci..
[6] Yann LeCun,et al. What is the best multi-stage architecture for object recognition? , 2009, 2009 IEEE 12th International Conference on Computer Vision.
[7] Zhenghao Chen,et al. On Random Weights and Unsupervised Feature Learning , 2011, ICML.
[8] Doina Precup,et al. An information-theoretic approach to curiosity-driven reinforcement learning , 2012, Theory in Biosciences.
[9] Pierre-Yves Oudeyer,et al. Exploration in Model-based Reinforcement Learning by Empirically Estimating Learning Progress , 2012, NIPS.
[10] Alex Graves,et al. Playing Atari with Deep Reinforcement Learning , 2013, ArXiv.
[11] Yoshua Bengio,et al. Learning Phrase Representations using RNN Encoder–Decoder for Statistical Machine Translation , 2014, EMNLP.
[12] Sergey Levine,et al. Incentivizing Exploration In Reinforcement Learning With Deep Predictive Models , 2015, ArXiv.
[13] Le Song,et al. Deep Fried Convnets , 2014, 2015 IEEE International Conference on Computer Vision (ICCV).
[14] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[15] Shane Legg,et al. Human-level control through deep reinforcement learning , 2015, Nature.
[16] Filip De Turck,et al. VIME: Variational Information Maximizing Exploration , 2016, NIPS.
[17] Benjamin Van Roy,et al. Deep Exploration via Bootstrapped DQN , 2016, NIPS.
[18] Alex Graves,et al. Asynchronous Methods for Deep Reinforcement Learning , 2016, ICML.
[19] Tom Schaul,et al. Unifying Count-Based Exploration and Intrinsic Motivation , 2016, NIPS.
[20] Demis Hassabis,et al. Mastering the game of Go with deep neural networks and tree search , 2016, Nature.
[21] Joshua B. Tenenbaum,et al. Hierarchical Deep Reinforcement Learning: Integrating Temporal Abstraction and Intrinsic Motivation , 2016, NIPS.
[22] Filip De Turck,et al. #Exploration: A Study of Count-Based Exploration for Deep Reinforcement Learning , 2016, NIPS.
[23] Marc G. Bellemare,et al. Count-Based Exploration with Neural Density Models , 2017, ICML.
[24] Alexei A. Efros,et al. Curiosity-Driven Exploration by Self-Supervised Prediction , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).
[25] Misha Denil,et al. Learning to Perform Physics Experiments via Deep Reinforcement Learning , 2016, ICLR.
[26] Quoc V. Le,et al. Neural Architecture Search with Reinforcement Learning , 2016, ICLR.
[27] Justin Fu,et al. EX2: Exploration with Exemplar Models for Deep Reinforcement Learning , 2017, NIPS.
[28] Daan Wierstra,et al. Variational Intrinsic Control , 2016, ICLR.
[29] Pieter Abbeel,et al. UCB and InfoGain Exploration via $\boldsymbol{Q}$-Ensembles , 2017, ArXiv.
[30] Marc G. Bellemare,et al. A Distributional Perspective on Reinforcement Learning , 2017, ICML.
[31] S. Shankar Sastry,et al. Surprise-Based Intrinsic Motivation for Deep Reinforcement Learning , 2017, ArXiv.
[32] Alec Radford,et al. Proximal Policy Optimization Algorithms , 2017, ArXiv.
[33] Tom Schaul,et al. Rainbow: Combining Improvements in Deep Reinforcement Learning , 2017, AAAI.
[34] Henryk Michalewski,et al. Expert-augmented actor-critic for ViZDoom and Montezumas Revenge , 2018, ArXiv.
[35] Marcin Andrychowicz,et al. Parameter Space Noise for Exploration , 2017, ICLR.
[36] David Budden,et al. Distributed Prioritized Experience Replay , 2018, ICLR.
[37] Daniel L. K. Yamins,et al. Learning to Play with Intrinsically-Motivated Self-Aware Agents , 2018, NeurIPS.
[38] Marlos C. Machado,et al. Revisiting the Arcade Learning Environment: Evaluation Protocols and Open Problems for General Agents , 2017, J. Artif. Intell. Res..
[39] Shane Legg,et al. Noisy Networks for Exploration , 2017, ICLR.
[40] Tim Salimans,et al. Learning Montezuma's Revenge from a Single Demonstration , 2018, ArXiv.
[41] Pieter Abbeel,et al. Variational Option Discovery Algorithms , 2018, ArXiv.
[42] Nando de Freitas,et al. Playing hard exploration games by watching YouTube , 2018, NeurIPS.
[43] Shane Legg,et al. IMPALA: Scalable Distributed Deep-RL with Importance Weighted Actor-Learner Architectures , 2018, ICML.
[44] Ilya Kostrikov,et al. Intrinsic Motivation and Automatic Curricula via Asymmetric Self-Play , 2017, ICLR.
[45] Sergey Levine,et al. Temporal Difference Models: Model-Free Deep RL for Model-Based Control , 2018, ICLR.
[46] Yonatan Loewenstein,et al. DORA The Explorer: Directed Outreaching Reinforcement Action-Selection , 2018, ICLR.
[47] Jeff Clune,et al. Deep Curiosity Search: Intra-Life Exploration Improves Performance on Challenging Deep Reinforcement Learning Problems , 2018, ArXiv.
[48] Rémi Munos,et al. Observe and Look Further: Achieving Consistent Performance on Atari , 2018, ArXiv.
[49] Ian Osband,et al. The Uncertainty Bellman Equation and Exploration , 2017, ICML.
[50] Albin Cassirer,et al. Randomized Prior Functions for Deep Reinforcement Learning , 2018, NeurIPS.
[51] Sergey Levine,et al. Diversity is All You Need: Learning Skills without a Reward Function , 2018, ICLR.
[52] Alexei A. Efros,et al. Large-Scale Study of Curiosity-Driven Learning , 2018, ICLR.
[53] Marlos C. Machado,et al. Count-Based Exploration with the Successor Representation , 2018, AAAI.
[54] Jakub W. Pachocki,et al. Learning dexterous in-hand manipulation , 2018, Int. J. Robotics Res..