Stochastic learning and optimization - A sensitivity-based approach
暂无分享,去创建一个
[1] A. F. Veinott. Discrete Dynamic Programming with Sensitive Discount Optimality Criteria , 1969 .
[2] W. Wonham. Linear Multivariable Control: A Geometric Approach , 1974 .
[3] Erhan Çinlar,et al. Introduction to stochastic processes , 1974 .
[4] C. D. Gelatt,et al. Optimization by Simulated Annealing , 1983, Science.
[5] Karl Johan Åström,et al. Adaptive Control , 1989, Embedded Digital Control with Microcontrollers.
[6] Xi-Ren Cao,et al. Perturbation analysis of discrete event dynamic systems , 1991 .
[7] Donald A. Sofge,et al. Handbook of Intelligent Control: Neural, Fuzzy, and Adaptive Approaches , 1992 .
[8] Lalit M. Patnaik,et al. Genetic algorithms: a survey , 1994, Computer.
[9] Xi-Ren Cao,et al. Realization Probabilities: The Dynamics of Queuing Systems , 1994 .
[10] Martin L. Puterman,et al. Markov Decision Processes: Discrete Stochastic Dynamic Programming , 1994 .
[11] Dimitri P. Bertsekas,et al. Dynamic Programming and Optimal Control, Two Volume Set , 1995 .
[12] Hans Zwart,et al. An Introduction to Infinite-Dimensional Linear Systems Theory , 1995, Texts in Applied Mathematics.
[13] John N. Tsitsiklis,et al. Neuro-Dynamic Programming , 1996, Encyclopedia of Machine Learning.
[14] Xi-Ren Cao,et al. Algorithms for sensitivity analysis of Markov systems through potentials and perturbation realization , 1998, IEEE Trans. Control. Syst. Technol..
[15] Christos G. Cassandras,et al. Introduction to Discrete Event Systems , 1999, The Kluwer International Series on Discrete Event Dynamic Systems.
[16] Leyuan Shi,et al. Nested Partitions Method for Global Optimization , 2000, Oper. Res..
[17] Peter L. Bartlett,et al. Infinite-Horizon Policy-Gradient Estimation , 2001, J. Artif. Intell. Res..
[18] John N. Tsitsiklis,et al. Simulation-based optimization of Markov reward processes , 2001, IEEE Trans. Autom. Control..
[19] Peter L. Bartlett,et al. Experiments with Infinite-Horizon, Policy-Gradient Estimation , 2001, J. Artif. Intell. Res..
[20] Paul Jung,et al. No free lunch. , 2002, Health affairs.
[21] Yu-Chi Ho,et al. The no free lunch theorems: complexity and security , 2003, IEEE Trans. Autom. Control..
[22] Christos G. Cassandras,et al. Perturbation analysis and control of two-class stochastic fluid models for communication networks , 2003, IEEE Trans. Autom. Control..
[23] Dirk P. Kroese,et al. The Cross-Entropy Method: A Unified Approach to Combinatorial Optimization, Monte-Carlo Simulation and Machine Learning , 2004 .
[24] Dirk P. Kroese,et al. The Cross Entropy Method: A Unified Approach To Combinatorial Optimization, Monte-carlo Simulation (Information Science and Statistics) , 2004 .
[25] Alexander B. Kurzhanski,et al. National Achievements in Control Theory (The Aerospace Perspective) , 2004 .
[26] Xi-Ren Cao,et al. A basic formula for online policy gradient algorithms , 2005, IEEE Transactions on Automatic Control.
[27] Richard S. Sutton,et al. Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.
[28] Michael C. Fu,et al. A Model Reference Adaptive Search Method for Global Optimization , 2007, Oper. Res..
[29] Y. Ho,et al. Ordinal Optimization: Soft Optimization for Hard Problems , 2007 .
[30] M. Fu. Perturbation Analysis , 2007 .
[31] Xi-Ren Cao,et al. Stochastic learning and optimization - A sensitivity-based approach , 2007, Annual Reviews in Control.