暂无分享,去创建一个
[1] Marcin Andrychowicz,et al. Multi-Goal Reinforcement Learning: Challenging Robotics Environments and Request for Research , 2018, ArXiv.
[2] Mohammad Norouzi,et al. Mastering Atari with Discrete World Models , 2020, ICLR.
[3] Pierre Baldi,et al. Curiosity-Driven Multi-Criteria Hindsight Experience Replay , 2019, ArXiv.
[4] Sergey Levine,et al. Model-Based Reinforcement Learning for Atari , 2019, ICLR.
[5] Demis Hassabis,et al. Mastering Atari, Go, chess and shogi by planning with a learned model , 2019, Nature.
[6] Mikael Henaff,et al. Explicit Explore-Exploit Algorithms in Continuous State Spaces , 2019, NeurIPS.
[7] Wojciech Jaskowski,et al. Model-Based Active Exploration , 2018, ICML.
[8] Pieter Abbeel,et al. Planning to Explore via Self-Supervised World Models , 2020, ICML.
[9] Ruben Villegas,et al. Learning Latent Dynamics for Planning from Pixels , 2018, ICML.
[10] Honglak Lee,et al. Sample-Efficient Reinforcement Learning with Stochastic Ensemble Value Expansion , 2018, NeurIPS.
[11] Jürgen Schmidhuber,et al. World Models , 2018, ArXiv.
[12] Amos J. Storkey,et al. Exploration by Random Network Distillation , 2018, ICLR.
[13] Tianhong Dai,et al. Episodic Self-Imitation Learning with Hindsight , 2020, Electronics.
[14] Lei Han,et al. Curriculum-guided Hindsight Experience Replay , 2019, NeurIPS.
[15] Shane Legg,et al. Human-level control through deep reinforcement learning , 2015, Nature.
[16] Sergey Levine,et al. Deep Reinforcement Learning in a Handful of Trials using Probabilistic Dynamics Models , 2018, NeurIPS.
[17] Jean-Baptiste Mouret,et al. Multi-objective Model-based Policy Search for Data-efficient Learning with Sparse Rewards , 2018, CoRL.
[18] Jakub W. Pachocki,et al. Learning dexterous in-hand manipulation , 2018, Int. J. Robotics Res..
[19] Giovanni Montana,et al. PlanGAN: Model-based Planning With Sparse Rewards and Multiple Goals , 2020, NeurIPS.
[20] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[21] Marcin Andrychowicz,et al. Hindsight Experience Replay , 2017, NIPS.
[22] Houqiang Li,et al. Soft Hindsight Experience Replay , 2020, ArXiv.
[23] Wojciech Zaremba,et al. OpenAI Gym , 2016, ArXiv.
[24] Pieter Abbeel,et al. Model-Ensemble Trust-Region Policy Optimization , 2018, ICLR.
[25] Sergey Levine,et al. Visual Reinforcement Learning with Imagined Goals , 2018, NeurIPS.
[26] Sergey Levine,et al. Deep Dynamics Models for Learning Dexterous Manipulation , 2019, CoRL.
[27] Alexei A. Efros,et al. Large-Scale Study of Curiosity-Driven Learning , 2018, ICLR.
[28] Yuval Tassa,et al. Data-efficient Deep Reinforcement Learning for Dexterous Manipulation , 2017, ArXiv.
[29] Jürgen Schmidhuber,et al. Formal Theory of Creativity, Fun, and Intrinsic Motivation (1990–2010) , 2010, IEEE Transactions on Autonomous Mental Development.
[30] Yuval Tassa,et al. Continuous control with deep reinforcement learning , 2015, ICLR.
[31] Alexei A. Efros,et al. Curiosity-Driven Exploration by Self-Supervised Prediction , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).
[32] Deepak Pathak,et al. Self-Supervised Exploration via Disagreement , 2019, ICML.
[33] Yoshua Bengio,et al. Revisiting Fundamentals of Experience Replay , 2020, ICML.
[34] Shuo Wang,et al. Curiosity-Driven Exploration for Off-Policy Reinforcement Learning Methods* , 2019, 2019 IEEE International Conference on Robotics and Biomimetics (ROBIO).
[35] Sergey Levine,et al. When to Trust Your Model: Model-Based Policy Optimization , 2019, NeurIPS.
[36] Sergey Levine,et al. Model-Based Value Estimation for Efficient Model-Free Reinforcement Learning , 2018, ArXiv.
[37] Richard S. Sutton,et al. Integrated Architectures for Learning, Planning, and Reacting Based on Approximating Dynamic Programming , 1990, ML.
[38] Jakub W. Pachocki,et al. Dota 2 with Large Scale Deep Reinforcement Learning , 2019, ArXiv.
[39] Pieter Abbeel,et al. Visual Hindsight Experience Replay , 2019, ArXiv.
[40] Tamim Asfour,et al. Model-Based Reinforcement Learning via Meta-Policy Optimization , 2018, CoRL.
[41] Richard S. Sutton,et al. A Deeper Look at Experience Replay , 2017, ArXiv.
[42] Sergey Levine,et al. Neural Network Dynamics for Model-Based Deep Reinforcement Learning with Model-Free Fine-Tuning , 2017, 2018 IEEE International Conference on Robotics and Automation (ICRA).
[43] Yuandong Tian,et al. Algorithmic Framework for Model-based Deep Reinforcement Learning with Theoretical Guarantees , 2018, ICLR.
[44] Yali Du,et al. MHER: Model-based Hindsight Experience Replay , 2021, ArXiv.
[45] Sergey Levine,et al. Continuous Deep Q-Learning with Model-based Acceleration , 2016, ICML.