暂无分享,去创建一个
[1] C. Bessaga. On the converse of Banach "fixed-point principle" , 1959 .
[2] Ronald A. Howard,et al. Dynamic Programming and Markov Processes , 1960 .
[3] Ruey-Wen Liu,et al. Construction of Suboptimal Control Sequences , 1967 .
[4] D. Kleinman. On an iterative technique for Riccati equation computations , 1968 .
[5] George N. Saridis,et al. An Approximation Theory of Optimal Control for Trainable Manipulators , 1979, IEEE Transactions on Systems, Man, and Cybernetics.
[6] Gerald B. Folland,et al. Real Analysis: Modern Techniques and Their Applications , 1984 .
[7] B. Anderson,et al. Optimal control: linear quadratic methods , 1990 .
[8] A. Bruckner,et al. Elementary Real Analysis , 1991 .
[9] Randal W. Beard,et al. Galerkin approximations of the generalized Hamilton-Jacobi-Bellman equation , 1997, Autom..
[10] Kenji Doya,et al. Reinforcement Learning in Continuous Time and Space , 2000, Neural Computation.
[11] W. A. Kirk,et al. Handbook of metric fixed point theory , 2001 .
[12] George G. Lendaris,et al. Adaptive dynamic programming , 2002, IEEE Trans. Syst. Man Cybern. Part C.
[13] Michail G. Lagoudakis,et al. Least-Squares Policy Iteration , 2003, J. Mach. Learn. Res..
[14] P. Loeb,et al. LUSIN'S THEOREM AND BOCHNER INTEGRATION , 2004, math/0406370.
[15] Richard S. Sutton,et al. Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.
[16] Frank L. Lewis,et al. Nearly optimal control laws for nonlinear systems with saturating actuators using a neural network HJB approach , 2005, Autom..
[17] Warren B. Powell,et al. Approximate Dynamic Programming - Solving the Curses of Dimensionality , 2007 .
[18] W. Haddad,et al. Nonlinear Dynamical Systems and Control: A Lyapunov-Based Approach , 2008 .
[19] Shie Mannor,et al. Regularized Policy Iteration , 2008, NIPS.
[20] Sean P. Meyn,et al. Q-learning and Pontryagin's Minimum Principle , 2009, Proceedings of the 48h IEEE Conference on Decision and Control (CDC) held jointly with 2009 28th Chinese Control Conference.
[21] Luigi Fortuna,et al. Reinforcement Learning and Adaptive Dynamic Programming for Feedback Control , 2009 .
[22] Frank L. Lewis,et al. 2009 Special Issue: Neural network approach to continuous-time direct adaptive optimal control for partially unknown nonlinear systems , 2009 .
[23] Shalabh Bhatnagar,et al. Toward Off-Policy Learning Control with Function Approximation , 2010, ICML.
[24] Jae Young Lee,et al. Integral Q-learning and explorized policy iteration for adaptive optimal control of continuous-time linear systems , 2012, Autom..
[25] Wulfram Gerstner,et al. Reinforcement Learning Using a Continuous Time Actor-Critic Framework with Spiking Neurons , 2013, PLoS Comput. Biol..
[26] F. Lewis,et al. Online adaptive algorithm for optimal control with integral reinforcement learning , 2014 .
[27] Jae Young Lee,et al. On integral generalized policy iteration for continuous-time linear quadratic regulations , 2014, Autom..
[28] Tingwen Huang,et al. Data-based approximate policy iteration for affine nonlinear continuous-time optimal control design , 2014, Autom..
[29] Jae Young Lee,et al. Integral Reinforcement Learning for Continuous-Time Input-Affine Nonlinear Systems With Simultaneous Invariant Explorations , 2015, IEEE Transactions on Neural Networks and Learning Systems.
[30] Frank L. Lewis,et al. Optimal Output-Feedback Control of Unknown Continuous-Time Linear Systems Using Off-policy Reinforcement Learning , 2016, IEEE Transactions on Cybernetics.
[31] David Barber,et al. Nesterov's accelerated gradient and momentum as approximations to regularised update descent , 2016, 2017 International Joint Conference on Neural Networks (IJCNN).