Learning Deep Robot Controllers by Exploiting Successful and Failed Executions
暂无分享,去创建一个
[1] Sergey Levine,et al. Learning deep control policies for autonomous aerial vehicles with MPC-guided policy search , 2015, 2016 IEEE International Conference on Robotics and Automation (ICRA).
[2] Kyungjae Lee,et al. Robust learning from demonstration using leveraged Gaussian processes and sparse-constrained optimization , 2016, 2016 IEEE International Conference on Robotics and Automation (ICRA).
[3] Carme Torras,et al. Dual REPS: A Generalization of Relative Entropy Policy Search Exploiting Bad Experiences , 2017, IEEE Transactions on Robotics.
[4] Shimon Whiteson,et al. Inverse Reinforcement Learning from Failure , 2016, AAMAS.
[5] Jan Peters,et al. Sample-based informationl-theoretic stochastic optimal control , 2014, 2014 IEEE International Conference on Robotics and Automation (ICRA).
[6] Yasemin Altun,et al. Relative Entropy Policy Search , 2010 .
[7] Sergey Levine,et al. Path integral guided policy search , 2016, 2017 IEEE International Conference on Robotics and Automation (ICRA).
[8] Aude Billard,et al. Learning from failed demonstrations in unreliable systems , 2013, 2013 13th IEEE-RAS International Conference on Humanoid Robots (Humanoids).
[9] Emanuel Todorov,et al. Combining the benefits of function approximation and trajectory optimization , 2014, Robotics: Science and Systems.
[10] Sergey Levine,et al. Guided Policy Search via Approximate Mirror Descent , 2016, NIPS.
[11] Jan Peters,et al. A Survey on Policy Search for Robotics , 2013, Found. Trends Robotics.
[12] Sergey Levine,et al. Learning Neural Network Policies with Guided Policy Search under Unknown Dynamics , 2014, NIPS.
[13] E. Todorov,et al. A generalized iterative LQG method for locally-optimal feedback control of constrained nonlinear stochastic systems , 2005, Proceedings of the 2005, American Control Conference, 2005..
[14] Sergey Levine,et al. Reset-free guided policy search: Efficient deep reinforcement learning with stochastic initial states , 2016, 2017 IEEE International Conference on Robotics and Automation (ICRA).
[15] Sergey Levine,et al. End-to-End Training of Deep Visuomotor Policies , 2015, J. Mach. Learn. Res..
[16] Yuval Tassa,et al. Synthesis and stabilization of complex behaviors through online trajectory optimization , 2012, 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems.
[17] Yuval Tassa,et al. Continuous control with deep reinforcement learning , 2015, ICLR.
[18] Aude Billard,et al. Robot Learning from Failed Demonstrations , 2012, Int. J. Soc. Robotics.
[19] Jan Peters,et al. Reinforcement learning in robotics: A survey , 2013, Int. J. Robotics Res..
[20] Jun Nakanishi,et al. Dynamical Movement Primitives: Learning Attractor Models for Motor Behaviors , 2013, Neural Computation.
[21] Nolan Wagener,et al. Learning contact-rich manipulation skills with guided policy search , 2015, 2015 IEEE International Conference on Robotics and Automation (ICRA).
[22] Gaurav S. Sukhatme,et al. Combining Model-Based and Model-Free Updates for Trajectory-Centric Reinforcement Learning , 2017, ICML.
[23] Sergey Levine,et al. Trust Region Policy Optimization , 2015, ICML.