A Heuristic Search Approach to Planning with Continuous Resources in Stochastic Domains

We consider the problem of optimal planning in stochastic domains with resource constraints, where the resources are continuous and the choice of action at each step depends on resource availability. We introduce the HAO* algorithm, a generalization of the AO* algorithm that performs search in a hybrid state space that is modeled using both discrete and continuous state variables, where the continuous variables represent monotonic resources. Like other heuristic search algorithms, HAO* leverages knowledge of the start state and an admissible heuristic to focus computational effort on those parts of the state space that could be reached from the start state by following an optimal policy. We show that this approach is especially effective when resource constraints limit how much of the state space is reachable. Experimental results demonstrate its effectiveness in the domain that motivates our research: automated planning for planetary exploration rovers.

[1]  Lihong Li,et al.  Lazy Approximation for Solving Continuous Finite-Horizon MDPs , 2005, AAAI.

[2]  Michael L. Littman,et al.  Exact Solutions to Time-Dependent MDPs , 2000, NIPS.

[3]  Judea Pearl,et al.  Heuristics : intelligent search strategies for computer problem solving , 1984 .

[4]  Subbarao Kambhampati,et al.  Effective Approaches for Partial Satisfaction (Over-Subscription) Planning , 2004, AAAI.

[5]  John N. Tsitsiklis,et al.  Neuro-Dynamic Programming , 1996, Encyclopedia of Machine Learning.

[6]  Shlomo Zilberstein,et al.  LAO*: A heuristic search algorithm that finds solutions with loops , 2001, Artif. Intell..

[7]  David E. Smith Choosing Objectives in Over-Subscription Planning , 2004, ICAPS.

[8]  Milind Tambe,et al.  A Fast Analytical Algorithm for Solving Markov Decision Processes with Real-Valued Resources , 2007, IJCAI.

[9]  John Rust Using Randomization to Break the Curse of Dimensionality , 1997 .

[10]  Nils J. Nilsson,et al.  Principles of Artificial Intelligence , 1980, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[11]  C. Guestrin,et al.  Solving Factored MDPs with Hybrid State and Action Variables , 2006, J. Artif. Intell. Res..

[12]  Ari K. Jónsson,et al.  Activity Planning for the Mars Exploration Rovers , 2005, ICAPS.

[13]  Andrew W. Moore,et al.  Variable Resolution Discretization in Optimal Control , 2002, Machine Learning.

[14]  Jesfis Peral,et al.  Heuristics -- intelligent search strategies for computer problem solving , 1984 .

[15]  Eugene Asarin,et al.  Widening the Boundary between Decidable and Undecidable Hybrid Systems , 2002, CONCUR.

[16]  Craig Boutilier,et al.  Decision-Theoretic Planning: Structural Assumptions and Computational Leverage , 1999, J. Artif. Intell. Res..

[17]  Carme Torras,et al.  An efficient algorithm for searching implicit AND/OR graphs with cycles , 2000, Artificial Intelligence.

[18]  E. Altman Constrained Markov Decision Processes , 1999 .

[19]  Ronen I. Brafman,et al.  Planning with Continuous Resources in Stochastic Domains , 2005, IJCAI.

[20]  R. Sargent,et al.  Mission planning and target tracking for autonomous instrument placement , 2005, 2005 IEEE Aerospace Conference.

[21]  Jon Louis Bentley,et al.  An Algorithm for Finding Best Matches in Logarithmic Expected Time , 1977, TOMS.

[22]  Ronen I. Brafman,et al.  Hierarchical Heuristic Forward Search in Stochastic Domains , 2007, IJCAI.

[23]  Martin L. Puterman,et al.  Markov Decision Processes: Discrete Stochastic Dynamic Programming , 1994 .

[24]  Zhengzhu Feng,et al.  Dynamic Programming for Structured Continuous Markov Decision Problems , 2004, UAI.

[25]  Ansgar Fehnker,et al.  Automotive Control Revisited: Linear Inequalities as Approximation of Reachable Sets , 1998, HSCC.

[26]  David E. Smith,et al.  Planning Under Continuous Time and Resource Uncertainty: A Challenge for AI , 2002, AIPS Workshop on Planning for Temporal Domains.

[27]  John K. Slaney,et al.  Decision-Theoretic Planning with non-Markovian Rewards , 2011, J. Artif. Intell. Res..