The Effects of Large Disturbances on On-Line Reinforcement Learning for aWalking Robot
暂无分享,去创建一个
Wouter Caarls | Pieter Jonker | Martijn Wisse | R. Babuska | E. Schuitema | Robert Babuška | M. Wisse | P. Jonker | E. Schuitema | W. Caarls
[1] Douglas C. Hittle,et al. Robust reinforcement learning control with static and dynamic stability , 2001 .
[2] Michael I. Jordan,et al. MASSACHUSETTS INSTITUTE OF TECHNOLOGY ARTIFICIAL INTELLIGENCE LABORATORY and CENTER FOR BIOLOGICAL AND COMPUTATIONAL LEARNING DEPARTMENT OF BRAIN AND COGNITIVE SCIENCES , 1996 .
[3] Jun Morimoto,et al. Robust Reinforcement Learning , 2005, Neural Computation.
[4] Roderic A. Grupen,et al. Robust Reinforcement Learning in Motion Planning , 1993, NIPS.
[5] Andrew W. Moore,et al. Locally Weighted Learning , 1997, Artificial Intelligence Review.
[6] Martijn Wisse,et al. The design of LEO: A 2D bipedal walking robot for online autonomous Reinforcement Learning , 2010, 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems.
[7] M. Coleman,et al. The simplest walking model: stability, complexity, and scaling. , 1998, Journal of biomechanical engineering.
[8] Dimitri P. Bertsekas,et al. Dynamic Programming: Deterministic and Stochastic Models , 1987 .
[9] Graham J. Williams,et al. On-Line Unsupervised Outlier Detection Using Finite Mixtures with Discounting Learning Algorithms , 2000, KDD '00.
[10] Richard S. Sutton,et al. Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.
[11] Douglas C. Hittle,et al. Robust Reinforcement Learning Control Using Integral Quadratic Constraints for Recurrent Neural Networks , 2007, IEEE Transactions on Neural Networks.
[12] John N. Tsitsiklis,et al. Asynchronous Stochastic Approximation and Q-Learning , 1994, Machine Learning.