Automatic Derivation of Finite-State Machines for Behavior Control

Finite-state controllers represent an effective action selection mechanisms widely used in domains such as video-games and mobile robotics. In contrast to the policies obtained from MDPs and POMDPs, finite-state controllers have two advantages: they are often extremely compact, and they are general, applying to many problems and not just one. A limitation of finite-state controllers, on the other hand, is that they are written by hand. In this paper, we address this limitation, presenting a method for deriving controllers automatically from models. The models represent a class of contingent problems where actions are deterministic and some fluents are observable. The problem of deriving a controller is converted into a conformant problem that is solved using classical planners, taking advantage of a complete translation into classical planning introduced recently. The controllers derived are 'general' in the sense that they do not solve the original problem only, but many variations as well, including changes in the size of the problem or in the uncertainty of the initial situation and action effects. Several experiments illustrating the automatic derivation of controllers are presented.

[1]  Hector Geffner,et al.  Compiling Uncertainty Away in Conformant Planning Problems with Bounded Width , 2009, J. Artif. Intell. Res..

[2]  Blai Bonet,et al.  Automatic Derivation of Memoryless Policies and Finite-State Controllers Using Classical Planners , 2009, ICAPS.

[3]  Michael L. Littman,et al.  Memoryless policies: theoretical limitations and practical results , 1994 .

[4]  Mat Buckland,et al.  Programming Game AI by Example , 2004 .

[5]  Rajesh P. N. Rao,et al.  Embodiment is the foundation, not a level , 1996, Behavioral and Brain Sciences.

[6]  Aude Billard,et al.  From Animals to Animats , 2004 .

[7]  David Chapman,et al.  Penguins Can Make Cake , 1989, AI Mag..

[8]  Kee-Eung Kim,et al.  Learning Finite-State Controllers for Partially Observable Environments , 1999, UAI.

[9]  Shlomo Zilberstein,et al.  Policy Iteration for Decentralized Control of Markov Decision Processes , 2009, J. Artif. Intell. Res..

[10]  Maja J. Mataric,et al.  The Robotics Primer , 2007 .

[11]  Robin R. Murphy,et al.  Introduction to AI Robotics , 2000 .

[12]  Leslie Pack Kaelbling,et al.  Planning and Acting in Partially Observable Stochastic Domains , 1998, Artif. Intell..

[13]  Craig Boutilier,et al.  Bounded Finite State Controllers , 2003, NIPS.

[14]  Hector J. Levesque,et al.  Planning with Loops , 2005, IJCAI.

[15]  Jonathan H. Connell,et al.  Minimalist mobile robotics - a colony-style architecture for an artificial creature , 1990, Perspectives in artificial intelligence.

[16]  Hector J. Levesque,et al.  What Is Planning in the Presence of Sensing? , 1996, AAAI/IAAI, Vol. 2.

[17]  Bernhard Nebel,et al.  In Defense of PDDL Axioms , 2003, IJCAI.