A Flexible Approach to Modeling Unpredictable Events in MDPs

In planning with a Markov decision process (MDP) framework, there is the implicit assumption that the world is predictable. Practitioners must simply take it on good faith the MDP they have constructed is comprehensive and accurate enough to model the exact probabilities with which all events may occur under all circumstances. Here, we challenge the conventional assumption of complete predictability, arguing that some events are inherently unpredictable. Towards more effectively modeling problems with unpredictable events, we develop a hybrid framework that explicitly distinguishes decision factors whose probabilities are not assigned precisely while still representing known probability components using conventional principled MDP transitions. Our approach is also flexible, resulting in a factored model of variable abstraction whose usage for planning results in different levels of approximation. We illustrate the application of our framework to an intelligent surveillance planning domain.

[1]  Gabriella Cortellessa,et al.  Automated planning and scheduling , 2014, Intelligenza Artificiale.

[2]  Leslie Pack Kaelbling,et al.  Influence-Based Abstraction for Multiagent Systems , 2012, AAAI.

[3]  Leslie Pack Kaelbling,et al.  Planning and Acting in Partially Observable Stochastic Domains , 1998, Artif. Intell..

[4]  Robert Givan,et al.  Bounded Parameter Markov Decision Processes , 1997, ECP.

[5]  Ran El-Yaniv,et al.  On Prediction Using Variable Order Markov Models , 2004, J. Artif. Intell. Res..

[6]  Judea Pearl,et al.  Probabilistic reasoning in intelligent systems - networks of plausible inference , 1991, Morgan Kaufmann series in representation and reasoning.

[7]  Patrick Fabiani,et al.  Approximate Policy Iteration for Generalized Semi-Markov Decision Processes: an Improved Algorithm , 2008 .

[8]  Håkan L. S. Younes,et al.  Solving Generalized Semi-Markov Decision Processes Using Continuous Phase-Type Distributions , 2004, AAAI.

[9]  Scott Sanner,et al.  Solutions to Factored MDPs with Imprecise Transition Probabilities 1 , 2011 .

[10]  S. Juneja,et al.  Rare-event Simulation Techniques : An Introduction and Recent Advances , 2006 .

[11]  Edmund H. Durfee,et al.  Flexible approximation of structured interactions in decentralized Markov decision processes , 2009, AAMAS.

[12]  D. Harmanec Generalizing Markov decision processes to imprecise probabilities , 2002 .

[13]  Victor R. Lesser,et al.  Decentralized Markov decision processes with event-driven interactions , 2004, Proceedings of the Third International Joint Conference on Autonomous Agents and Multiagent Systems, 2004. AAMAS 2004..

[14]  Richard S. Sutton,et al.  Predictive Representations of State , 2001, NIPS.

[15]  P. Shahabuddin,et al.  Chapter 11 Rare-Event Simulation Techniques: An Introduction and Recent Advances , 2006, Simulation.

[16]  Craig Boutilier,et al.  Decision-Theoretic Planning: Structural Assumptions and Computational Leverage , 1999, J. Artif. Intell. Res..

[17]  Gerardo Rubino,et al.  Rare Event Simulation using Monte Carlo Methods , 2009 .

[18]  Zhi-Qiang Liu,et al.  Bounded-Parameter Partially Observable Markov Decision Processes , 2008, ICAPS.

[19]  Edmund H. Durfee,et al.  Influence-Based Policy Abstraction for Weakly-Coupled Dec-POMDPs , 2010, ICAPS.

[20]  Xi-Ren Cao,et al.  Event-Based Optimization of Markov Systems , 2008, IEEE Transactions on Automatic Control.

[21]  R. Goldman,et al.  "Unrolling" Complex Task Models into MDPs , 2007, AAAI Spring Symposium: Game Theoretic and Decision Theoretic Agents.

[22]  Martin L. Puterman,et al.  Markov Decision Processes: Discrete Stochastic Dynamic Programming , 1994 .

[23]  Rina Dechter,et al.  Improving Bound Propagation , 2006, ECAI.

[24]  Hideaki Itoh,et al.  Partially observable Markov decision processes with imprecise parameters , 2007, Artif. Intell..