Speeding Up Planning in Markov Decision Processes via Automatically Constructed Abstraction
暂无分享,去创建一个
Vadim Bulitko | Russell Greiner | Csaba Szepesvári | Alejandro Isaza | Csaba Szepesvari | A. Isaza | V. Bulitko | R. Greiner
[1] Eleanor Clark,et al. Baldur's Gate , 1970 .
[2] John N. Tsitsiklis,et al. Neuro-Dynamic Programming , 1996, Encyclopedia of Machine Learning.
[3] Robert Givan,et al. Model Reduction Techniques for Computing Approximately Optimal Solutions for Markov Decision Processes , 1997, UAI.
[4] Milos Hauskrecht,et al. Hierarchical Solution of Markov Decision Processes using Macro-actions , 1998, UAI.
[5] IT Kee-EungKim. Solving Factored MDPs Using Non-homogeneous Partitions , 1998 .
[6] Doina Precup,et al. Between MDPs and Semi-MDPs: A Framework for Temporal Abstraction in Reinforcement Learning , 1999, Artif. Intell..
[7] N. Carpenter,et al. WarCraft III: reign of chaos , 2003, SVR '03.
[8] Manfred Huber,et al. State Space Reduction For Hierarchical Reinforcement Learning , 2004, FLAIRS.
[9] Geoffrey J. Gordon,et al. Fast Exact Planning in Markov Decision Processes , 2005, ICAPS.
[10] Nathan R. Sturtevant,et al. Memory-Efficient Abstractions for Pathfinding , 2007, AIIDE.
[11] Manfred Huber,et al. Effective Control Knowledge Transfer through Learning Skill and Representation Hierarchies , 2007, IJCAI.
[12] Jonathan Schaeffer,et al. Dynamic Control in Path-Planning with Real-Time Heuristic Search , 2007, ICAPS.
[13] Nathan R. Sturtevant,et al. Graph Abstraction in Real-time Heuristic Search , 2007, J. Artif. Intell. Res..
[14] Jonathan Schaeffer,et al. Dynamic Control in Real-Time Heuristic Search , 2008, J. Artif. Intell. Res..