暂无分享,去创建一个
Pieter Abbeel | David Held | Xinyang Geng | Carlos Florensa | P. Abbeel | David Held | Carlos Florensa | Xinyang Geng
[1] Jürgen Schmidhuber,et al. Curious model-building control systems , 1991, [Proceedings] 1991 IEEE International Joint Conference on Neural Networks.
[2] Andrew Y. Ng,et al. Policy Invariance Under Reward Transformations: Theory and Application to Reward Shaping , 1999, ICML.
[3] Jason Weston,et al. Curriculum learning , 2009, ICML '09.
[4] Jürgen Schmidhuber,et al. Formal Theory of Creativity, Fun, and Intrinsic Motivation (1990–2010) , 2010, IEEE Transactions on Autonomous Mental Development.
[5] Daphne Koller,et al. Self-Paced Learning for Latent Variable Models , 2010, NIPS.
[6] Pierre-Yves Oudeyer,et al. Intrinsically motivated goal exploration for active motor learning in robots: A case study , 2010, 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems.
[7] Andrew G. Barto,et al. Intrinsically Motivated Hierarchical Skill Learning in Structured Environments , 2010, IEEE Transactions on Autonomous Mental Development.
[8] Scott Kuindersma,et al. Autonomous Skill Acquisition on a Mobile Manipulator , 2011, AAAI.
[9] Michiel van de Panne,et al. Curriculum Learning for Motor Skills , 2012, Canadian Conference on AI.
[10] Yuval Tassa,et al. MuJoCo: A physics engine for model-based control , 2012, 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems.
[11] Ruslan Salakhutdinov,et al. Learning Stochastic Feedforward Neural Networks , 2013, NIPS.
[12] Pierre-Yves Oudeyer,et al. Active learning of inverse models with intrinsically motivated goal exploration in robots , 2013, Robotics Auton. Syst..
[13] Jan Peters,et al. A Survey on Policy Search for Robotics , 2013, Found. Trends Robotics.
[14] Jan Peters,et al. Data-Efficient Generalization of Robot Skills with Contextual Policy Search , 2013, AAAI.
[15] Wojciech Zaremba,et al. Learning to Execute , 2014, ArXiv.
[16] Yoshua Bengio,et al. Generative Adversarial Nets , 2014, NIPS.
[17] Alexander Fabisch,et al. Active contextual policy search , 2014, J. Mach. Learn. Res..
[18] Peter Englert,et al. Multi-task policy search for robotics , 2014, 2014 IEEE International Conference on Robotics and Automation (ICRA).
[19] Shiguang Shan,et al. Self-Paced Curriculum Learning , 2015, AAAI.
[20] Tom Schaul,et al. Universal Value Function Approximators , 2015, ICML.
[21] Sergey Levine,et al. Trust Region Policy Optimization , 2015, ICML.
[22] Pravesh Ranchod,et al. Nonparametric Bayesian reward segmentation for skill discovery using inverse reinforcement learning , 2015, 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).
[23] Samy Bengio,et al. Scheduled Sampling for Sequence Prediction with Recurrent Neural Networks , 2015, NIPS.
[24] Shane Legg,et al. Human-level control through deep reinforcement learning , 2015, Nature.
[25] Yuval Tassa,et al. Continuous control with deep reinforcement learning , 2015, ICLR.
[26] Pieter Abbeel,et al. Benchmarking Deep Reinforcement Learning for Continuous Control , 2016, ICML.
[27] Filip De Turck,et al. VIME: Variational Information Maximizing Exploration , 2016, NIPS.
[28] Yisong Yue,et al. Generating Long-term Trajectories Using Deep Hierarchical Networks , 2016, NIPS.
[29] Alex Graves,et al. Strategic Attentive Writer for Learning Macro-Actions , 2016, NIPS.
[30] Tom Schaul,et al. Unifying Count-Based Exploration and Intrinsic Motivation , 2016, NIPS.
[31] Pieter Abbeel,et al. Value Iteration Networks , 2016, NIPS.
[32] Demis Hassabis,et al. Mastering the game of Go with deep neural networks and tree search , 2016, Nature.
[33] Honglak Lee,et al. Deep Learning for Reward Design to Improve Monte Carlo Tree Search in ATARI Games , 2016, IJCAI.
[34] Sergey Levine,et al. End-to-End Training of Deep Visuomotor Policies , 2015, J. Mach. Learn. Res..
[35] J. Schulman,et al. Variational Information Maximizing Exploration , 2016 .
[36] Sergey Levine,et al. High-Dimensional Continuous Control Using Generalized Advantage Estimation , 2015, ICLR.
[37] Yuval Tassa,et al. Learning and Transfer of Modulated Locomotor Controllers , 2016, ArXiv.
[38] Balaraman Ravindran,et al. Online Multi-Task Learning Using Biased Sampling , 2017 .
[39] Filip De Turck,et al. #Exploration: A Study of Count-Based Exploration for Deep Reinforcement Learning , 2016, NIPS.
[40] Alex Graves,et al. Automated Curriculum Learning for Neural Networks , 2017, ICML.
[41] Pieter Abbeel,et al. Stochastic Neural Networks for Hierarchical Reinforcement Learning , 2016, ICLR.
[42] Marcin Andrychowicz,et al. Hindsight Experience Replay , 2017, NIPS.
[43] Raymond Y. K. Lau,et al. Least Squares Generative Adversarial Networks , 2016, 2017 IEEE International Conference on Computer Vision (ICCV).
[44] Pieter Abbeel,et al. Reverse Curriculum Generation for Reinforcement Learning , 2017, CoRL.
[45] S. Shankar Sastry,et al. Surprise-Based Intrinsic Motivation for Deep Reinforcement Learning , 2017, ArXiv.
[46] Balaraman Ravindran,et al. Learning to Multi-Task by Active Sampling , 2017, ICLR.
[47] Ilya Kostrikov,et al. Intrinsic Motivation and Automatic Curricula via Asymmetric Self-Play , 2017, ICLR.
[48] Zhen Wang,et al. On the Effectiveness of Least Squares Generative Adversarial Networks , 2017, IEEE Transactions on Pattern Analysis and Machine Intelligence.