暂无分享,去创建一个
Honglak Lee | Junhyuk Oh | Mohammad Norouzi | Jongwook Choi | Marcin Moczulski | Yijie Guo | Neal Wu | Junhyuk Oh | Honglak Lee | Mohammad Norouzi | Yijie Guo | Jongwook Choi | Marcin Moczulski | Neal Wu
[1] William M. Rand,et al. Objective Criteria for the Evaluation of Clustering Methods , 1971 .
[2] F. Baeyens,et al. Contingency awareness in evaluative conditioning: A case for unaware affective-evaluative learning , 1990 .
[3] J. Urgen Schmidhuber,et al. Adaptive confidence and adaptive curiosity , 1991, Forschungsberichte, TU Munich.
[4] Moses Charikar,et al. Similarity estimation techniques from rounding algorithms , 2002, STOC '02.
[5] Nuttapong Chentanez,et al. Intrinsically Motivated Reinforcement Learning , 2004, NIPS.
[6] Hugh F. Durrant-Whyte,et al. Simultaneous localization and mapping: part I , 2006, IEEE Robotics & Automation Magazine.
[7] Pierre-Yves Oudeyer,et al. What is Intrinsic Motivation? A Typology of Computational Approaches , 2007, Frontiers Neurorobotics.
[8] Michael L. Littman,et al. An analysis of model-based Interval Estimation for Markov Decision Processes , 2008, J. Comput. Syst. Sci..
[9] Marc G. Bellemare,et al. Investigating Contingency Awareness Using Atari 2600 Games , 2012, AAAI.
[10] Michael I. Jordan,et al. Revisiting k-means: New Algorithms via Bayesian Nonparametrics , 2011, ICML.
[11] Andrew G. Barto,et al. Intrinsic Motivation and Reinforcement Learning , 2013, Intrinsically Motivated Learning in Natural and Artificial Systems.
[12] David C Rowland,et al. Place cells, grid cells, and memory. , 2015, Cold Spring Harbor perspectives in biology.
[13] Yoshua Bengio,et al. Show, Attend and Tell: Neural Image Caption Generation with Visual Attention , 2015, ICML.
[14] Shane Legg,et al. Human-level control through deep reinforcement learning , 2015, Nature.
[15] Honglak Lee,et al. Action-Conditional Video Prediction using Deep Networks in Atari Games , 2015, NIPS.
[16] Marc G. Bellemare,et al. The Arcade Learning Environment: An Evaluation Platform for General Agents , 2012, J. Artif. Intell. Res..
[17] Yoshua Bengio,et al. Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.
[18] Filip De Turck,et al. VIME: Variational Information Maximizing Exploration , 2016, NIPS.
[19] Benjamin Van Roy,et al. Deep Exploration via Bootstrapped DQN , 2016, NIPS.
[20] Alex Graves,et al. Asynchronous Methods for Deep Reinforcement Learning , 2016, ICML.
[21] Tom Schaul,et al. Unifying Count-Based Exploration and Intrinsic Motivation , 2016, NIPS.
[22] Jitendra Malik,et al. Learning to Poke by Poking: Experiential Learning of Intuitive Physics , 2016, NIPS.
[23] Ramón Fernández Astudillo,et al. From Softmax to Sparsemax: A Sparse Model of Attention and Multi-Label Classification , 2016, ICML.
[24] J. Watson,et al. THE DEVELOPMENT AND GENERALIZATION OF "CONTINGENCY AWARENESS" IN EARLY INFANCY: SOME HYPOTHESES , 2016 .
[25] Marcus Hutter,et al. Count-Based Exploration in Feature Space for Reinforcement Learning , 2017, IJCAI.
[26] Filip De Turck,et al. #Exploration: A Study of Count-Based Exploration for Deep Reinforcement Learning , 2016, NIPS.
[27] Alexei A. Efros,et al. Curiosity-Driven Exploration by Self-Supervised Prediction , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).
[28] Joelle Pineau,et al. Independently Controllable Features , 2017 .
[29] A. P. Hyper-parameters. Count-Based Exploration with Neural Density Models , 2017 .
[30] Trevor Darrell,et al. Loss is its own Reward: Self-Supervision for Reinforcement Learning , 2016, ICLR.
[31] Razvan Pascanu,et al. Learning to Navigate in Complex Environments , 2016, ICLR.
[32] S. Shankar Sastry,et al. Surprise-Based Intrinsic Motivation for Deep Reinforcement Learning , 2017, ArXiv.
[33] Tom Schaul,et al. Reinforcement Learning with Unsupervised Auxiliary Tasks , 2016, ICLR.
[34] Alec Radford,et al. Proximal Policy Optimization Algorithms , 2017, ArXiv.
[35] Yoshihide Sawada,et al. Disentangling Controllable and Uncontrollable Factors of Variation by Interacting with the World , 2018, ArXiv.
[36] Marcin Andrychowicz,et al. Parameter Space Noise for Exploration , 2017, ICLR.
[37] Satinder Singh,et al. On Learning Intrinsic Rewards for Policy Gradient Methods , 2018, NeurIPS.
[38] Marlos C. Machado,et al. Revisiting the Arcade Learning Environment: Evaluation Protocols and Open Problems for General Agents , 2017, J. Artif. Intell. Res..
[39] Raia Hadsell,et al. Learning to Navigate in Cities Without a Map , 2018, NeurIPS.
[40] Amos J. Storkey,et al. Exploration by Random Network Distillation , 2018, ICLR.
[41] Sergey Levine,et al. Diversity is All You Need: Learning Skills without a Reward Function , 2018, ICLR.
[42] Murray Shanahan,et al. Feature Control as Intrinsic Motivation for Hierarchical Reinforcement Learning , 2017, IEEE Transactions on Neural Networks and Learning Systems.
[43] Alexei A. Efros,et al. Large-Scale Study of Curiosity-Driven Learning , 2018, ICLR.