Learning context conditions for BDI plan selection

An important drawback to the popular Belief, Desire, and Intentions (BDI) paradigm is that such systems include no element of learning from experience. In particular, the so-called context conditions of plans, on which the whole model relies for plan selection, are restricted to be boolean formulas that are to be specified at design/implementation time. To address these limitations, we propose a novel BDI programming framework that, by suitably modeling context conditions as decision trees, allows agents to learn the probability of success for plans based on previous execution experiences. By using a probabilistic plan selection function, the agents can balance exploration and exploitation of their plans. We develop and empirically investigate two extreme approaches to learning the new context conditions and show that both can be advantageous in certain situations. Finally, we propose a generalization of the probabilistic plan selection function that yields a middle-ground between the two extreme approaches, and which we thus argue is the most flexible and simple approach.

[1]  Michael Winikoff,et al.  Avoiding Resource Conflicts in Intelligent Agents , 2002, ECAI.

[2]  Hector Muñoz-Avila,et al.  Learning HTN Method Preconditions and Action Models from Partial Observations , 2009, IJCAI.

[3]  Ian P. W. Sillitoe,et al.  A Fast Memory-Efficient Incremental Decision Tree Algorithm in its Application to Mobile Robot Navigation , 2006, 2006 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[4]  David J. Israel,et al.  Plans and resource‐bounded practical reasoning , 1988, Comput. Intell..

[5]  Paul E. Utgoff,et al.  Decision Tree Induction Based on Efficient Tree Restructuring , 1997, Machine Learning.

[6]  Mehdi Dastani,et al.  2APL: a practical agent programming language , 2008, Autonomous Agents and Multi-Agent Systems.

[7]  Anand S. Rao,et al.  AgentSpeak(L): BDI Agents Speak Out in a Logical Computable Language , 1996, MAAMAW.

[8]  Michael Wooldridge,et al.  Programming Multi-Agent Systems in AgentSpeak using Jason (Wiley Series in Agent Technology) , 2007 .

[9]  Martin A. Riedmiller,et al.  Karlsruhe Brainstormers - A Reinforcement Learning Approach to Robotic Soccer , 2000, RoboCup.

[10]  Karen L. Myers,et al.  The SPARK agent framework , 2004, Proceedings of the Third International Joint Conference on Autonomous Agents and Multiagent Systems, 2004. AAMAS 2004..

[11]  Steve S. Benfield,et al.  Making a strong business case for multiagent technology , 2006, AAMAS '06.

[12]  Michael P. Georgeff,et al.  Decision-Making in an Embedded Reasoning System , 1989, IJCAI.

[13]  Lin Padgham,et al.  Enhancing the Adaptation of BDI Agents Using Learning Techniques , 2009, Int. J. Agent Technol. Syst..

[14]  Amal El Fallah Seghrouchni,et al.  Learning in BDI Multi-agent Systems , 2004, CLIMA.

[15]  Ian Witten,et al.  Data Mining , 2000 .

[16]  Koen V. Hindriks,et al.  Agent Programming in 3APL , 1999, Autonomous Agents and Multi-Agent Systems.

[17]  Martha E. Pollack,et al.  The Uses of Plans , 1992, Artif. Intell..

[18]  Paolo Busetta,et al.  Jack intelligent agents - components for intelligent agents in java , 1998 .

[19]  Thomas G. Dietterich What is machine learning? , 2020, Archives of Disease in Childhood.

[20]  Ian H. Witten,et al.  Data mining: practical machine learning tools and techniques with Java implementations , 2002, SGMD.