Generalized Policy Iteration for continuous-time systems
暂无分享,去创建一个
[1] K. N. Dollman,et al. - 1 , 1743 .
[2] Paul J. Werbos,et al. Approximate dynamic programming for real-time control and neural modeling , 1992 .
[3] F. Lewis,et al. Model-free Q-learning designs for discrete-time zero-sum games with application to H-infinity control , 2007, 2007 European Control Conference (ECC).
[4] Victor M. Becerra,et al. Optimal control , 2008, Scholarpedia.
[5] Lyle Noakes,et al. Continuous-Time Adaptive Critics , 2007, IEEE Transactions on Neural Networks.
[6] Andrew G. Barto,et al. Adaptive linear quadratic control using policy iteration , 1994, Proceedings of 1994 American Control Conference - ACC '94.
[7] P. Werbos,et al. Beyond Regression : "New Tools for Prediction and Analysis in the Behavioral Sciences , 1974 .
[8] Frank L. Lewis,et al. Policy Iterations on the Hamilton–Jacobi–Isaacs Equation for $H_{\infty}$ State Feedback Control With Input Saturation , 2006, IEEE Transactions on Automatic Control.
[9] John N. Tsitsiklis,et al. On the Convergence of Optimistic Policy Iteration , 2002, J. Mach. Learn. Res..
[10] George G. Lendaris,et al. Adaptive dynamic programming , 2002, IEEE Trans. Syst. Man Cybern. Part C.
[11] Frank L. Lewis,et al. Model-free Q-learning designs for linear discrete-time zero-sum games with application to H-infinity control , 2007, Autom..
[12] Richard S. Sutton,et al. Introduction to Reinforcement Learning , 1998 .
[13] Frank L. Lewis,et al. Aircraft Control and Simulation , 1992 .
[14] Ruey-Wen Liu,et al. Construction of Suboptimal Control Sequences , 1967 .
[15] Paul J. Werbos,et al. Neural networks for control and system identification , 1989, Proceedings of the 28th IEEE Conference on Decision and Control,.
[16] Tomas Landelius,et al. Reinforcement Learning and Distributed Local Model Synthesis , 1997 .
[17] Warren B. Powell,et al. Handbook of Learning and Approximate Dynamic Programming , 2006, IEEE Transactions on Automatic Control.
[18] Frank L. Lewis,et al. Adaptive optimal control for continuous-time linear systems based on policy iteration , 2009, Autom..
[19] John N. Tsitsiklis,et al. Neuro-Dynamic Programming , 1996, Encyclopedia of Machine Learning.
[20] Randal W. Beard,et al. Galerkin approximations of the generalized Hamilton-Jacobi-Bellman equation , 1997, Autom..
[21] Richard S. Sutton,et al. Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.
[22] Frank L. Lewis,et al. Adaptive Critic Designs for Discrete-Time Zero-Sum Games With Application to $H_{\infty}$ Control , 2007, IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics).
[23] Frank L. Lewis,et al. Nearly optimal control laws for nonlinear systems with saturating actuators using a neural network HJB approach , 2005, Autom..
[24] Kurt Hornik,et al. Universal approximation of an unknown mapping and its derivatives using multilayer feedforward networks , 1990, Neural Networks.
[25] D. Kleinman. On an iterative technique for Riccati equation computations , 1968 .
[26] Kenji Doya,et al. Reinforcement Learning in Continuous Time and Space , 2000, Neural Computation.
[27] Frank L. Lewis,et al. Adaptive optimal control algorithm for continuous-time nonlinear systems based on policy iteration , 2008, 2008 47th IEEE Conference on Decision and Control.