Learning Complex Neural Network Policies with Trajectory Optimization
暂无分享,去创建一个
[1] A. Laub,et al. Generalized eigenproblem algorithms and software for algebraic Riccati equations , 1984, Proceedings of the IEEE.
[2] Max Donath,et al. American Control Conference , 1993 .
[3] Jun Nakanishi,et al. Learning Attractor Landscapes for Learning Motor Primitives , 2002, NIPS.
[4] E. Todorov,et al. A generalized iterative LQG method for locally-optimal feedback control of constrained nonlinear stochastic systems , 2005, Proceedings of the 2005, American Control Conference, 2005..
[5] Stephen P. Boyd,et al. Convex Optimization , 2004, Algorithms and Theory of Computation Handbook.
[6] Emanuel Todorov,et al. Linearly-solvable Markov decision problems , 2006, NIPS.
[7] KangKang Yin,et al. SIMBICON: simple biped locomotion control , 2007, ACM Trans. Graph..
[8] Stefan Schaal,et al. Applying the Episodic Natural Actor-Critic Architecture to Motor Primitive Learning , 2007, ESANN.
[9] M. Giles. An extended collection of matrix derivative results for forward and reverse mode algorithmic dieren tiation , 2008 .
[10] Stefan Schaal,et al. 2008 Special Issue: Reinforcement learning of motor skills with policy gradients , 2008 .
[11] Jan Peters,et al. Learning motor primitives for robotics , 2009, 2009 IEEE International Conference on Robotics and Automation.
[12] Stefan Schaal,et al. Reinforcement learning of motor skills in high dimensions: A path integral approach , 2010, 2010 IEEE International Conference on Robotics and Automation.
[13] J. Andrew Bagnell,et al. Modeling Purposeful Adaptive Behavior with the Principle of Maximum Causal Entropy , 2010 .
[14] Carl E. Rasmussen,et al. PILCO: A Model-Based and Data-Efficient Approach to Policy Search , 2011, ICML.
[15] Geoffrey J. Gordon,et al. A Reduction of Imitation Learning and Structured Prediction to No-Regret Online Learning , 2010, AISTATS.
[16] J. Zico Kolter,et al. Design, analysis, and learning control of a fully actuated micro wind turbine , 2012, 2012 American Control Conference (ACC).
[17] J. Andrew Bagnell,et al. Agnostic System Identification for Model-Based Reinforcement Learning , 2012, ICML.
[18] Marc Toussaint,et al. On Stochastic Optimal Control and Reinforcement Learning by Approximate Inference , 2012, Robotics: Science and Systems.
[19] Vicenç Gómez,et al. Optimal control as a graphical model inference problem , 2009, Machine Learning.
[20] Yuval Tassa,et al. MuJoCo: A physics engine for model-based control , 2012, 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems.
[21] Martial Hebert,et al. Learning monocular reactive UAV control in cluttered natural environments , 2012, 2013 IEEE International Conference on Robotics and Automation.
[22] Sergey Levine,et al. Exploring Deep and Recurrent Architectures for Optimal Control , 2013, ArXiv.
[23] Sergey Levine,et al. Guided Policy Search , 2013, ICML.
[24] Jan Peters,et al. Probabilistic Movement Primitives , 2013, NIPS.
[25] Jan Peters,et al. Reinforcement learning in robotics: A survey , 2013, Int. J. Robotics Res..
[26] Jan Peters,et al. A Survey on Policy Search for Robotics , 2013, Found. Trends Robotics.
[27] Sergey Levine,et al. Variational Policy Search via Trajectory Optimization , 2013, NIPS.
[28] Emanuel Todorov,et al. Combining the benefits of function approximation and trajectory optimization , 2014, Robotics: Science and Systems.