Kernelized value function approximation for reinforcement learning
暂无分享,去创建一个
[1] D. Bertsekas,et al. Adaptive aggregation methods for infinite horizon dynamic programming , 1989 .
[2] Justin A. Boyan,et al. Least-Squares Temporal Difference Learning , 1999, ICML.
[3] C. Rasmussen,et al. Gaussian Process Priors with Uncertain Inputs - Application to Multiple-Step Ahead Time Series Forecasting , 2002, NIPS.
[4] Shie Mannor,et al. Sparse Online Greedy Support Vector Regression , 2002, ECML.
[5] Carl E. Rasmussen,et al. Gaussian Processes in Reinforcement Learning , 2003, NIPS.
[6] Michail G. Lagoudakis,et al. Reinforcement Learning as Classification: Leveraging Modern Classifiers , 2003, ICML.
[7] Shie Mannor,et al. Reinforcement learning with Gaussian processes , 2005, ICML.
[8] Christopher M. Bishop,et al. Pattern Recognition and Machine Learning (Information Science and Statistics) , 2006 .
[9] Alborz Geramifard,et al. Incremental Least-Squares Temporal Difference Learning , 2006, AAAI.
[10] Xin Xu,et al. Kernel Least-Squares Temporal Difference Learning , 2006 .
[11] Xin Xu,et al. Kernel-Based Least Squares Policy Iteration for Reinforcement Learning , 2007, IEEE Transactions on Neural Networks.
[12] Radford M. Neal. Pattern Recognition and Machine Learning , 2007, Technometrics.
[13] Sridhar Mahadevan,et al. Proto-value Functions: A Laplacian Framework for Learning Representation and Control in Markov Decision Processes , 2007, J. Mach. Learn. Res..
[14] Lihong Li,et al. An analysis of linear models, linear value-function approximation, and feature selection for reinforcement learning , 2008, ICML '08.
[15] Shie Mannor,et al. Regularized Policy Iteration , 2008, NIPS.