暂无分享,去创建一个
Weinan Zhang | Yong Yu | Jian Shen | Hang Lai
[1] Erik Talvitie,et al. Self-Correcting Models for Model-Based Reinforcement Learning , 2016, AAAI.
[2] Nikolai Matni,et al. On the Sample Complexity of the Linear Quadratic Regulator , 2017, Foundations of Computational Mathematics.
[3] Csaba Szepesvári,et al. Model-based reinforcement learning with nearly tight exploration complexity bounds , 2010, ICML.
[4] Richard S. Sutton,et al. Dyna, an integrated architecture for learning, planning, and reacting , 1990, SGAR.
[5] Pieter Abbeel,et al. Model-Ensemble Trust-Region Policy Optimization , 2018, ICLR.
[6] C. Rasmussen,et al. Improving PILCO with Bayesian Neural Network Dynamics Models , 2016 .
[7] Sergey Levine,et al. Recall Traces: Backtracking Models for Efficient Reinforcement Learning , 2018, ICLR.
[8] Jimmy Ba,et al. Exploring Model-based Planning with Policy Networks , 2019, ICLR.
[9] Kavosh Asadi,et al. Lipschitz Continuity in Model-based Reinforcement Learning , 2018, ICML.
[10] Yuval Tassa,et al. MuJoCo: A physics engine for model-based control , 2012, 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems.
[11] Sergey Levine,et al. Optimal control with learned local models: Application to dexterous manipulation , 2016, 2016 IEEE International Conference on Robotics and Automation (ICRA).
[12] Martial Hebert,et al. Improving Multi-Step Prediction of Learned Time Series Models , 2015, AAAI.
[13] Alec Radford,et al. Proximal Policy Optimization Algorithms , 2017, ArXiv.
[14] Nhat M. Nguyen. Improving model-based RL with Adaptive Rollout using Uncertainty Estimation , 2018 .
[15] Yuandong Tian,et al. Algorithmic Framework for Model-based Deep Reinforcement Learning with Theoretical Guarantees , 2018, ICLR.
[16] Shane Legg,et al. Human-level control through deep reinforcement learning , 2015, Nature.
[17] Pieter Abbeel,et al. Benchmarking Model-Based Reinforcement Learning , 2019, ArXiv.
[18] Luke Metz,et al. Learning to Predict Without Looking Ahead: World Models Without Forward Prediction , 2019, NeurIPS.
[19] Dieter Fox,et al. Gaussian Processes and Reinforcement Learning for Identification and Control of an Autonomous Blimp , 2007, Proceedings 2007 IEEE International Conference on Robotics and Automation.
[20] Pieter Abbeel,et al. Prediction and Control with Temporal Segment Models , 2017, ICML.
[21] Balaraman Ravindran,et al. EPOpt: Learning Robust Neural Network Policies Using Model Ensembles , 2016, ICLR.
[22] Yifan Wu,et al. Learning to Combat Compounding-Error in Model-Based Reinforcement Learning , 2019, ArXiv.
[23] Sergey Levine,et al. Learning Neural Network Policies with Guided Policy Search under Unknown Dynamics , 2014, NIPS.
[24] Ashley D. Edwards,et al. Forward-Backward Reinforcement Learning , 2018, ArXiv.
[25] Simon Osindero,et al. Conditional Generative Adversarial Nets , 2014, ArXiv.
[26] Kavosh Asadi,et al. Combating the Compounding-Error Problem with a Multi-step Model , 2019, ArXiv.
[27] Richard S. Sutton,et al. Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.
[28] Rob Fergus,et al. Understanding the Asymptotic Performance of Model-Based RL Methods , 2018 .
[29] Alborz Geramifard,et al. Dyna-Style Planning with Linear Function Approximation and Prioritized Sweeping , 2008, UAI.
[30] Jan Peters,et al. A Survey on Policy Search for Robotics , 2013, Found. Trends Robotics.
[31] Lihong Li,et al. An analysis of linear models, linear value-function approximation, and feature selection for reinforcement learning , 2008, ICML '08.
[32] Sebastian Engell,et al. Model Predictive Control Using Neural Networks [25 Years Ago] , 1995, IEEE Control Systems.
[33] Sergey Levine,et al. Model-Based Reinforcement Learning for Atari , 2019, ICLR.
[34] Sergey Levine,et al. When to Trust Your Model: Model-Based Policy Optimization , 2019, NeurIPS.
[35] Carl E. Rasmussen,et al. PILCO: A Model-Based and Data-Efficient Approach to Policy Search , 2011, ICML.
[36] Sergey Levine,et al. Soft Actor-Critic: Off-Policy Maximum Entropy Deep Reinforcement Learning with a Stochastic Actor , 2018, ICML.
[37] Sergey Levine,et al. Deep Reinforcement Learning in a Handful of Trials using Probabilistic Dynamics Models , 2018, NeurIPS.
[38] Sergey Levine,et al. Guided Policy Search , 2013, ICML.
[39] Byron Boots,et al. Dual Policy Iteration , 2018, NeurIPS.
[40] Sergey Levine,et al. Neural Network Dynamics for Model-Based Deep Reinforcement Learning with Model-Free Fine-Tuning , 2017, 2018 IEEE International Conference on Robotics and Automation (ICRA).
[41] Ian J. Goodfellow,et al. NIPS 2016 Tutorial: Generative Adversarial Networks , 2016, ArXiv.
[42] Marko Bacic,et al. Model predictive control , 2003 .
[43] Yuval Tassa,et al. Continuous control with deep reinforcement learning , 2015, ICLR.
[44] Carl E. Rasmussen,et al. Gaussian Processes in Reinforcement Learning , 2003, NIPS.
[45] Kavosh Asadi,et al. Towards a Simple Approach to Multi-step Model-based Reinforcement Learning , 2018, ArXiv.
[46] Hao Su,et al. Model Imitation for Model-Based Reinforcement Learning , 2019, ArXiv.