Semi-online neural-Q/spl I.bar/leaming for real-time robot learning
暂无分享,去创建一个
[1] Richard S. Sutton,et al. Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.
[2] G. Tesauro. Practical Issues in Temporal Difference Learning , 1992 .
[3] RamAshwin,et al. Experiments with reinforcement learning in problems with continuous state and action spaces , 1998 .
[4] Richard S. Sutton,et al. Introduction to Reinforcement Learning , 1998 .
[5] Andrew W. Moore,et al. Variable Resolution Dynamic Programming , 1991, ML Workshop.
[6] Leemon C. Baird,et al. Residual Algorithms: Reinforcement Learning with Function Approximation , 1995, ICML.
[7] Thomas G. Dietterich. What is machine learning? , 2020, Archives of Disease in Childhood.
[8] Manuela M. Veloso,et al. Tree Based Discretization for Continuous State Space Reinforcement Learning , 1998, AAAI/IAAI.
[9] Simon Haykin,et al. Neural Networks: A Comprehensive Foundation , 1998 .
[10] Chris Gaskett,et al. Q-Learning for Robot Control , 2002 .
[11] Pere Ridao,et al. Vision-based localization of an underwater robot in a structured environment , 2003, 2003 IEEE International Conference on Robotics and Automation (Cat. No.03CH37422).
[12] Leslie Pack Kaelbling,et al. Making Reinforcement Learning Work on Real Robots , 2002 .
[13] Richard S. Sutton,et al. Learning to predict by the methods of temporal differences , 1988, Machine Learning.
[14] Marios M. Polycarpou,et al. An analytical framework for local feedforward networks , 1998, IEEE Trans. Neural Networks.
[15] Peter Dayan,et al. Q-learning , 1992, Machine Learning.