Action rule induction from cause-effect pairs learned through robot-teacher interaction

In this work we propose a decision-making system that efficiently learns behaviors in the form of rules using natural human instructions about cause-effect relations in currently observed situations, avoiding complicated instructions and explanations of long-run action sequences and complete world dynamics. The learned rules are represented in a way suitable to both reactive and deliberative approaches, which are thus smoothly integrated. Simple and repetitive tasks are resolved reactively, while complex tasks would be faced in a more deliberative manner using a planner module. Human interaction is only required if the system fails to obtain the expected results when applying a rule, or fails to resolve the task with the knowledge acquired so far. I. INTRODUCTION N this work we are facing the problem of decision making for a multitask robot embedded in a human environment that should rapidly learn to perform tasks by interacting with humans, in an on-line way, and without any previous knowledge of the world dynamics or the tasks to be performed. From a very general point of view, we must consider two alternative approaches to the goal of building an intelligent agent: the deliberative and the reactive approaches. The deliberative approach began with the very birth of AI, and it is based on the principle of rationality (1), which states that "If an agent has knowledge that one of its actions will lead to one of its goals, then the agent will select that action.". The proponents of the knowledge-based systems using the principle of rationality soon realized that there are a number of important shortcomings with this approach, ranging from the frame problem (2), the difficulty of building a large enough database of knowledge providing the grounds for common sense, and the theorems stating the complexity of planning for even some of the simplest kinds of logical problems.

[1]  John McCarthy,et al.  SOME PHILOSOPHICAL PROBLEMS FROM THE STANDPOINT OF ARTI CIAL INTELLIGENCE , 1987 .

[2]  Allen Newell,et al.  The Knowledge Level , 1989, Artif. Intell..

[3]  Marcel Schoppers,et al.  Universal Plans for Reactive Robots in Unpredictable Environments , 1987, IJCAI.

[4]  R. A. Brooks,et al.  Intelligence without Representation , 1991, Artif. Intell..

[5]  Xuemei Wang,et al.  Learning Planning Operators by Observation and Practice , 1994, AIPS.

[6]  Scott Benson,et al.  Inductive Learning of Reactive Action Models , 1995, ICML.

[7]  Paul R. Cohen,et al.  Learning Planning Operators with Conditional and Probabilistic Effects , 1996 .

[8]  E. Gat On Three-Layer Architectures , 1997 .

[9]  Monica N. Nicolescu,et al.  A hierarchical architecture for behavior-based robots , 2002, AAMAS '02.

[10]  Richard S. Sutton,et al.  Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.

[11]  Christopher W. Geib,et al.  Object Action Complexes as an Interface for Planning and Robot Control , 2006 .

[12]  S. Harnad Symbol grounding problem , 1991, Scholarpedia.

[13]  John Levine,et al.  Evolving macro-actions for planning , 2007 .

[14]  É. Belin,et al.  Interaction between reactive and deliberative tasks for on-line decision-making , 2007 .

[15]  Dong-Ning Rao,et al.  Learning Activation Rules for Derived Predicates from Plan Examples , 2007 .

[16]  S. Yoon Towards Model-lite Planning : A Proposal For Learning & Planning with Incomplete Domain Models , 2007 .