暂无分享,去创建一个
Brendan McCane | Haitao Xu | Craig Atkinson | Lech Szymanski | B. McCane | Haitao Xu | Lech Szymanski | C. Atkinson
[1] Marek Wydmuch,et al. ViZDoom Competitions: Playing Doom From Pixels , 2018, IEEE Transactions on Games.
[2] Sergey Levine,et al. Incentivizing Exploration In Reinforcement Learning With Deep Predictive Models , 2015, ArXiv.
[3] Pieter Abbeel,et al. Benchmarking Deep Reinforcement Learning for Continuous Control , 2016, ICML.
[4] Filip De Turck,et al. VIME: Variational Information Maximizing Exploration , 2016, NIPS.
[5] Sergey Levine,et al. Trust Region Policy Optimization , 2015, ICML.
[6] Demis Hassabis,et al. Mastering the game of Go with deep neural networks and tree search , 2016, Nature.
[7] Tom Schaul,et al. Unifying Count-Based Exploration and Intrinsic Motivation , 2016, NIPS.
[8] Alexei A. Efros,et al. Large-Scale Study of Curiosity-Driven Learning , 2018, ICLR.
[9] Jürgen Schmidhuber,et al. Curious model-building control systems , 1991, [Proceedings] 1991 IEEE International Joint Conference on Neural Networks.
[10] Shane Legg,et al. Human-level control through deep reinforcement learning , 2015, Nature.
[11] Alex Graves,et al. Playing Atari with Deep Reinforcement Learning , 2013, ArXiv.
[12] Marc G. Bellemare,et al. Count-Based Exploration with Neural Density Models , 2017, ICML.
[13] Jürgen Schmidhuber,et al. Optimal Artificial Curiosity, Creativity, Music, and the Fine Arts , 2005 .
[14] Alexei A. Efros,et al. Curiosity-Driven Exploration by Self-Supervised Prediction , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).
[15] S. Shankar Sastry,et al. Surprise-Based Intrinsic Motivation for Deep Reinforcement Learning , 2017, ArXiv.
[16] Tom Schaul,et al. Rainbow: Combining Improvements in Deep Reinforcement Learning , 2017, AAAI.
[17] Frank L. Lewis,et al. Neural-network approximation of piecewise continuous functions: application to friction compensation , 2002, IEEE Trans. Neural Networks.
[18] Myron Tribus,et al. Thermostatics and thermodynamics : an introduction to energy, information and states of matter, with engineering applications , 1961 .
[19] Jürgen Schmidhuber,et al. A possibility for implementing curiosity and boredom in model-building neural controllers , 1991 .
[20] Ralph Linsker,et al. Self-organization in a perceptual network , 1988, Computer.
[21] Juergen Schmidhuber,et al. Unsupervised Minimax: Adversarial Curiosity, Generative Adversarial Networks, and Predictability Minimization , 2019, ArXiv.
[22] Pierre-Yves Oudeyer,et al. Exploration in Model-based Reinforcement Learning by Empirically Estimating Learning Progress , 2012, NIPS.
[23] Pierre-Yves Oudeyer,et al. Intrinsic Motivation Systems for Autonomous Mental Development , 2007, IEEE Transactions on Evolutionary Computation.
[24] Brendan McCane,et al. VASE: Variational Assorted Surprise Exploration for Reinforcement Learning , 2019, IEEE Transactions on Neural Networks and Learning Systems.
[25] Alec Radford,et al. Proximal Policy Optimization Algorithms , 2017, ArXiv.
[26] Amos J. Storkey,et al. Exploration by Random Network Distillation , 2018, ICLR.
[27] Terrence J. Sejnowski,et al. An Information-Maximization Approach to Blind Separation and Blind Deconvolution , 1995, Neural Computation.