A goal- and dependency-directed algorithm for learning hierarchical task networks

This paper describes a system for learning domain-dependent knowledge in the form of goal-indexed Hierarchical Task Networks (HTNs). DLIGHT is a goal-directed incremental learning algorithm which observes solution traces and generates rules for solving problems. One of the main challenges in learning this kind of knowledge is determining a good level of generality. Analytical methods, such as explanation-based macro-operator learning, construct very specific structures that guarantee a successful execution when applicable but generalize poorly to new problems. Previous goal-directed learning approaches produce hierarchical rules with more relaxed preconditions, but the learned knowledge suffers from over-generality. Our approach builds on one such approach but it strikes a better balance between generality and specificity. This is done by carrying out a goal-dependency analysis to determine the structure of the hierarchy and precondition of each rule to follow the successful solutions more closely while maintaining generality. We hypothesize that this algorithm produces HTNs that generalize well and can solve problems efficiently. We evaluate the system's behavior experimentally in several planning scenarios and conclude with related work and future research paths.

[1]  Dana S. Nau,et al.  SHOP2: An HTN Planning System , 2003, J. Artif. Intell. Res..

[2]  Tom M. Mitchell,et al.  LEAP: A Learning Apprentice for VLSI Design , 1985, IJCAI.

[3]  Jonathan Schaeffer,et al.  Learning Partial-Order Macros from Solutions , 2005, ICAPS.

[4]  John E. Laird,et al.  Learning Goal Hierarchies from Structured Observations and Expert Annotations , 2004, ILP.

[5]  David W. Aha,et al.  Learning approximate preconditions for methods in hierarchical plans , 2005, ICML.

[6]  Pat Langley,et al.  Learning hierarchical task networks by observation , 2006, ICML.

[7]  Dana S. Nau,et al.  Complexity results for hierarchical task-network planning , 1996 .

[8]  Earl D. Sacerdoti,et al.  The Nonlinear Nature of Plans , 1975, IJCAI.

[9]  Richard E. Korf,et al.  Planning as Search: A Quantitative Approach , 1987, Artif. Intell..

[10]  Claude Sammut,et al.  Automatic construction of reactive control systems using symbolic machine learning , 1996, The Knowledge Engineering Review.

[11]  Henry Lieberman,et al.  Watch what I do: programming by demonstration , 1993 .

[12]  Prasad Tadepalli,et al.  Learning Goal-Decomposition Rules Using Exercises , 1997, AAAI/IAAI.

[13]  Raymond J. Mooney,et al.  Generalizing the Order of Operators in Macro-Operators , 1988, ML.

[14]  Hector Muñoz-Avila,et al.  HTN-MAKER: Learning HTNs with Minimal Additional Knowledge Engineering Required , 2008, AAAI.

[15]  S. Griffis EDITOR , 1997, Journal of Navigation.

[16]  Ugur Kuter,et al.  Inductive Generalization of Analytically Learned Goal Hierarchies , 2009, ILP.

[17]  Ke Xu,et al.  A Domain-Independent System for Case-Based Task Decomposition without Domain Theories , 2005, AAAI.