A Partially-Observable Markov Decision Process for Dealing with Dynamically Changing Environments
暂无分享,去创建一个
[1] Joelle Pineau,et al. Online Planning Algorithms for POMDPs , 2008, J. Artif. Intell. Res..
[2] Joelle Pineau,et al. Bayesian reinforcement learning in continuous POMDPs with application to robot navigation , 2008, 2008 IEEE International Conference on Robotics and Automation.
[3] Leslie Pack Kaelbling,et al. Approximate Planning in POMDPs with Macro-Actions , 2003, NIPS.
[4] Joelle Pineau,et al. Active Learning in Partially Observable Markov Decision Processes , 2005, ECML.
[5] Jesse Hoey,et al. An analytic solution to discrete Bayesian reinforcement learning , 2006, ICML.
[6] Lancelot F. James,et al. Gibbs Sampling Methods for Stick-Breaking Priors , 2001 .
[7] R. Kohn,et al. On Gibbs sampling for state space models , 1994 .
[8] Joelle Pineau,et al. Bayes-Adaptive POMDPs , 2007, NIPS.
[9] David B. Dunson,et al. The dynamic hierarchical Dirichlet process , 2008, ICML '08.
[10] Lonnie Chrisman,et al. Reinforcement Learning with Perceptual Aliasing: The Perceptual Distinctions Approach , 1992, AAAI.
[11] D. Siegmund. Importance Sampling in the Monte Carlo Study of Sequential Tests , 1976 .
[12] Michael I. Jordan,et al. Hierarchical Dirichlet Processes , 2006 .
[13] Guy Shani,et al. Noname manuscript No. (will be inserted by the editor) A Survey of Point-Based POMDP Solvers , 2022 .
[14] Leslie Pack Kaelbling,et al. Learning Policies for Partially Observable Environments: Scaling Up , 1997, ICML.
[15] Finale Doshi-Velez,et al. The Infinite Partially Observable Markov Decision Process , 2009, NIPS.
[16] Joelle Pineau,et al. Bayes-Adaptive POMDPs: A New Perspective on the Explore-Exploit Tradeoff in Partially Observable Domains , 2008, ISAIM.