State Aggregation in Monte Carlo Tree Search
暂无分享,去创建一个
[1] Carmel Domshlak,et al. Friends or Foes? An AI Planning Perspective on Abstraction and Search , 2006, ICAPS.
[2] Thomas J. Walsh,et al. Integrating Sample-Based Planning and Model-Based Reinforcement Learning , 2010, AAAI.
[3] Yishay Mansour,et al. A Sparse Sampling Algorithm for Near-Optimal Planning in Large Markov Decision Processes , 1999, Machine Learning.
[4] Simon M. Lucas,et al. A Survey of Monte Carlo Tree Search Methods , 2012, IEEE Transactions on Computational Intelligence and AI in Games.
[5] Alan Fern,et al. UCT for Tactical Assault Planning in Real-Time Strategy Games , 2009, IJCAI.
[6] Benjamin Van Roy. Performance Loss Bounds for Approximate Value Iteration with State Aggregation , 2006, Math. Oper. Res..
[7] Thomas J. Walsh,et al. Towards a Unified Theory of State Abstraction for MDPs , 2006, AI&M.
[8] Doina Precup,et al. Metrics for Finite Markov Decision Processes , 2004, AAAI.
[9] David Silver,et al. Combining online and offline knowledge in UCT , 2007, ICML '07.
[10] Csaba Szepesvári,et al. Bandit Based Monte-Carlo Planning , 2006, ECML.
[11] Nan Jiang,et al. Improving UCT planning via approximate homomorphisms , 2014, AAMAS.
[12] Robert Givan,et al. Equivalence notions and model minimization in Markov decision processes , 2003, Artif. Intell..