Using supervised training signals of observable state dynamics to speed-up and improve reinforcement learning
暂无分享,去创建一个
[1] Charles W. Anderson,et al. Learning and problem-solving with multilayer connectionist systems (adaptive, strategy learning, neural networks, reinforcement learning) , 1986 .
[2] Richard S. Sutton,et al. Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.
[3] Satinder P. Singh,et al. On discovery and learning of models with predictive representations of state for agents with continuous actions and observations , 2007, AAMAS '07.
[4] Joerg Blasius,et al. Geometric Data Analysis , 2000 .
[5] Carl E. Rasmussen,et al. Model-Based Reinforcement Learning with Continuous States and Actions , 2008, ESANN.
[6] Mitsuo Kawato,et al. Multiple Model-Based Reinforcement Learning , 2002, Neural Computation.
[7] Darwin G. Caldwell,et al. Comparative Evaluation of Reinforcement Learning with Scalar Rewards and Linear Regression with Multidimensional Feedback , 2013 .
[8] Brigitte Le Roux,et al. Geometric Data Analysis , 2005 .
[9] Haibo He,et al. Goal Representation Heuristic Dynamic Programming on Maze Navigation , 2013, IEEE Transactions on Neural Networks and Learning Systems.