暂无分享,去创建一个
[1] Sergey Levine,et al. Temporal Difference Models: Model-Free Deep RL for Model-Based Control , 2018, ICLR.
[2] Tom Schaul,et al. Successor Features for Transfer in Reinforcement Learning , 2016, NIPS.
[3] Shane Legg,et al. Human-level control through deep reinforcement learning , 2015, Nature.
[4] Alex Graves,et al. Asynchronous Methods for Deep Reinforcement Learning , 2016, ICML.
[5] Richard S. Sutton,et al. Generalization in Reinforcement Learning: Successful Examples Using Sparse Coarse Coding , 1995, NIPS.
[6] Peter Stone,et al. Transfer Learning for Reinforcement Learning Domains: A Survey , 2009, J. Mach. Learn. Res..
[7] Sergey Levine,et al. End-to-End Training of Deep Visuomotor Policies , 2015, J. Mach. Learn. Res..
[8] Tom Schaul,et al. Reinforcement Learning with Unsupervised Auxiliary Tasks , 2016, ICLR.
[9] Marcin Andrychowicz,et al. Multi-Goal Reinforcement Learning: Challenging Robotics Environments and Request for Research , 2018, ArXiv.
[10] Yuval Tassa,et al. Continuous control with deep reinforcement learning , 2015, ICLR.
[11] Yuval Tassa,et al. MuJoCo: A physics engine for model-based control , 2012, 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems.
[12] Tom Schaul,et al. Transfer in Deep Reinforcement Learning Using Successor Features and Generalised Policy Improvement , 2018, ICML.
[13] Peter Dayan,et al. Improving Generalization for Temporal Difference Learning: The Successor Representation , 1993, Neural Computation.
[14] Rich Caruana,et al. Multitask Learning , 1997, Machine-mediated learning.
[15] Wojciech Zaremba,et al. OpenAI Gym , 2016, ArXiv.
[16] Samuel Gershman,et al. Deep Successor Reinforcement Learning , 2016, ArXiv.
[17] Tom Schaul,et al. Universal Value Function Approximators , 2015, ICML.
[18] Marcin Andrychowicz,et al. Hindsight Experience Replay , 2017, NIPS.
[19] Patrick M. Pilarski,et al. Horde: a scalable real-time architecture for learning knowledge from unsupervised sensorimotor interaction , 2011, AAMAS.