Refining Incomplete Planning Domain Models Through Plan Traces

Most existing work on learning planning models assumes that the entire model needs to be learned from scratch. A more realistic situation is that the planning agent has an incomplete model which it needs to refine through learning. In this paper we propose and evaluate a method for doing this. Our method takes as input an incomplete model (with missing preconditions and effects in the actions), as well as a set of plan traces that are known to be correct. It outputs a "refined" model that not only captures additional precondition/effect knowledge about the given actions, but also "macro actions". We use a MAX-SAT framework for learning, where the constraints are derived from the executability of the given plan traces, as well as the preconditions/ effects of the given incomplete model. Unlike traditional macro-action learners which use macros to increase the efficiency of planning (in the context of a complete model), our motivation for learning macros is to increase the accuracy (robustness) of the plans generated with the refined model. We demonstrate the effectiveness of our approach through a systematic empirical evaluation.

[1]  Piergiorgio Bertoli,et al.  Web Service Composition as Planning, Revisited: In Between Background Theories and Initial State Uncertainty , 2007, AAAI.

[2]  Richard Fikes,et al.  Learning and Executing Generalized Robot Plans , 1993, Artif. Intell..

[3]  Kristian J. Hammond,et al.  Case-Based Planning: Viewing Planning as a Memory Task , 1989 .

[4]  Felip Manyà,et al.  New Inference Rules for Max-SAT , 2007, J. Artif. Intell. Res..

[5]  Qiang Yang,et al.  Learning complex action models with quantifiers and logical implications , 2010, Artif. Intell..

[6]  Jianyong Wang,et al.  Mining sequential patterns by pattern-growth: the PrefixSpan approach , 2004, IEEE Transactions on Knowledge and Data Engineering.

[7]  Piergiorgio Bertoli,et al.  Automated composition of Web services via planning in asynchronous domains , 2005, Artif. Intell..

[8]  Subbarao Kambhampati,et al.  Synthesizing Robust Plans under Incomplete Domain Models , 2011, NIPS.

[9]  Richard Alterman,et al.  An Adaptive Planner , 1986, AAAI.

[10]  Andrew Garland,et al.  Plan evaluation with incomplete action descriptions , 2002, AAAI/IAAI.

[11]  Subbarao Kambhampati,et al.  Model-lite Planning for the Web Age Masses: The Challenges of Planning with Incomplete and Evolving Domain Models , 2007, AAAI.

[12]  Leslie Pack Kaelbling,et al.  Learning Planning Rules in Noisy Stochastic Worlds , 2005, AAAI.

[13]  Yolanda Gil,et al.  Automatically composed workflows for grid environments , 2004, IEEE Intelligent Systems.

[14]  Mohammed J. Zaki,et al.  SPADE: An Efficient Algorithm for Mining Frequent Sequences , 2004, Machine Learning.

[15]  Nils J. Nilsson,et al.  Artificial Intelligence , 1974, IFIP Congress.

[16]  Subbarao Kambhampati,et al.  Model-Lite Case-Based Planning , 2012, AAAI.

[17]  Richard E. Korf,et al.  Artificial intelligence journal special issue on heuristic search , 1999 .

[18]  Jonathan Schaeffer,et al.  Macro-FF: Improving AI Planning with Automatically Learned Macro-Operators , 2005, J. Artif. Intell. Res..

[19]  John Levine,et al.  Learning Macro-Actions for Arbitrary Planners and Domains , 2007, ICAPS.

[20]  Thomas G. Dietterich What is machine learning? , 2020, Archives of Disease in Childhood.

[21]  Richard E. Korf,et al.  Macro-Operators: A Weak Method for Learning , 1985, Artif. Intell..

[22]  Glenn A. Iba,et al.  A Heuristic Approach to the Discovery of Macro-Operators , 1989, Machine Learning.

[23]  Qiang Yang,et al.  Learning action models from plan examples using weighted MAX-SAT , 2007, Artif. Intell..

[24]  Andrew Coles,et al.  Marvin: A Heuristic Search Planner with Online Macro-Action Learning , 2011, J. Artif. Intell. Res..

[25]  Daniel Bryce,et al.  Planning and Acting in Incomplete Domains , 2011, ICAPS.