Acting Optimally in Partially Observable Stochastic Domains
暂无分享,去创建一个
Leslie Pack Kaelbling | Michael L. Littman | Anthony R. Cassandra | M. Littman | A. Cassandra | L. Kaelbling
[1] R. Bellman,et al. Dynamic Programming and Markov Processes , 1960 .
[2] Karl Johan Åström,et al. Optimal control of Markov processes with incomplete state information , 1965 .
[3] Edward J. Sondik,et al. The optimal control of par-tially observable Markov processes , 1971 .
[4] Edward J. Sondik,et al. The Optimal Control of Partially Observable Markov Processes over a Finite Horizon , 1973, Oper. Res..
[5] Robert C. Moore. A Formal Theory of Knowledge and Action , 1984 .
[6] John N. Tsitsiklis,et al. The Complexity of Markov Decision Processes , 1987, Math. Oper. Res..
[7] Richard S. Sutton,et al. Integrated Architectures for Learning, Planning, and Reacting Based on Approximating Dynamic Programming , 1990, ML.
[8] Ming Tan,et al. Cost-Sensitive Reinforcement Learning for Adaptive Classification and Control , 1991, AAAI.
[9] W. Lovejoy. A survey of algorithmic methods for partially observed Markov decision processes , 1991 .
[10] Lonnie Chrisman,et al. Reinforcement Learning with Perceptual Aliasing: The Perceptual Distinctions Approach , 1992, AAAI.
[11] Leslie Pack Kaelbling,et al. Planning With Deadlines in Stochastic Domains , 1993, AAAI.
[12] Andrew McCallum,et al. Overcoming Incomplete Perception with Utile Distinction Memory , 1993, ICML.
[13] Thomas G. Dietterich. What is machine learning? , 2020, Archives of Disease in Childhood.