暂无分享,去创建一个
[1] Brett Browning,et al. A survey of robot learning from demonstration , 2009, Robotics Auton. Syst..
[2] Sergey Levine,et al. Neural Network Dynamics for Model-Based Deep Reinforcement Learning with Model-Free Fine-Tuning , 2017, 2018 IEEE International Conference on Robotics and Automation (ICRA).
[3] Todd D. Murphey,et al. Model-Based Control Using Koopman Operators , 2017, Robotics: Science and Systems.
[4] Sergey Levine,et al. MBMF: Model-Based Priors for Model-Free Reinforcement Learning , 2017, ArXiv.
[5] Prabhat Nagarajan,et al. Learning Latent State Spaces for Planning through Reward Prediction , 2019, ArXiv.
[6] Todd D. Murphey,et al. Active Learning of Dynamics for Data-Driven Control Using Koopman Operators , 2019, IEEE Transactions on Robotics.
[7] Todd D. Murphey,et al. Sequential Action Control: Closed-Form Optimal Control for Nonlinear and Nonsmooth Systems , 2016, IEEE Transactions on Robotics.
[8] Sanjoy Dasgupta,et al. Off-Policy Temporal Difference Learning with Function Approximation , 2001, ICML.
[9] Carl E. Rasmussen,et al. PILCO: A Model-Based and Data-Efficient Approach to Policy Search , 2011, ICML.
[10] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[11] Nolan Wagener,et al. Information theoretic MPC for model-based reinforcement learning , 2017, 2017 IEEE International Conference on Robotics and Automation (ICRA).
[12] Dieter Fox,et al. Model-Based Generalization Under Parameter Uncertainty Using Path Integral Control , 2020, IEEE Robotics and Automation Letters.
[13] James M. Rehg,et al. Aggressive driving with model predictive path integral control , 2016, 2016 IEEE International Conference on Robotics and Automation (ICRA).
[14] Yishay Mansour,et al. Policy Gradient Methods for Reinforcement Learning with Function Approximation , 1999, NIPS.
[15] Sergey Levine,et al. Dynamics-Aware Unsupervised Discovery of Skills , 2019, ICLR.
[16] Evangelos Theodorou,et al. Relative entropy and free energy dualities: Connections to Path Integral and KL control , 2012, 2012 IEEE 51st IEEE Conference on Decision and Control (CDC).
[17] Sergey Levine,et al. Path integral guided policy search , 2016, 2017 IEEE International Conference on Robotics and Automation (ICRA).
[18] Alec Radford,et al. Proximal Policy Optimization Algorithms , 2017, ArXiv.
[19] Alexei A. Efros,et al. Curiosity-Driven Exploration by Self-Supervised Prediction , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).
[20] Erik I. Verriest,et al. Gradient Descent Approach to Optimal Mode Scheduling in Hybrid Dynamical Systems , 2008 .
[21] J. Andrew Bagnell,et al. Efficient Reductions for Imitation Learning , 2010, AISTATS.
[22] Dean Pomerleau,et al. ALVINN, an autonomous land vehicle in a neural network , 2015 .
[23] Justin A. Boyan,et al. Least-Squares Temporal Difference Learning , 1999, ICML.
[24] Emanuel Todorov,et al. Iterative Linear Quadratic Regulator Design for Nonlinear Biological Movement Systems , 2004, ICINCO.
[25] Sergey Levine,et al. Soft Actor-Critic: Off-Policy Maximum Entropy Deep Reinforcement Learning with a Stochastic Actor , 2018, ICML.
[26] Sergey Levine,et al. Learning Neural Network Policies with Guided Policy Search under Unknown Dynamics , 2014, NIPS.
[27] Sergey Levine,et al. Deep Reinforcement Learning in a Handful of Trials using Probabilistic Dynamics Models , 2018, NeurIPS.