Multiagent Expedition with Graphical Models

We investigate a class of multiagent planning problems termed multiagent expedition, where agents move around an open, unknown, partially observable, stochastic, and physical environment, in pursuit of multiple and alternative goals of different utility. Optimal planning in multiagent expedition is highly intractable. We introduce the notion of conditional optimality, decompose the task into a set of semi-independent optimization subtasks, and apply a decision-theoretic multiagent graphical model to solve each subtask optimally. A set of techniques are proposed to enhance modeling so that the resultant graphical model can be practically evaluated. Effectiveness of the framework and its scalability are demonstrated through experiments. Multiagent expedition can be characterized as decentralized partially observable Markov decision processes (Dec-POMDPs). Hence, this work contributes towards practical planning in Dec-POMDPs.

[1]  Andreas Birk,et al.  Autonomous Mapping in the Real Robots Rescue League , 2003 .

[2]  Yang Xiang,et al.  A Decision-Theoretic Graphical Model for Collaborative Design on Supply Chains , 2004, Canadian Conference on AI.

[3]  James Kennedy,et al.  Particle swarm optimization , 1995, Proceedings of ICNN'95 - International Conference on Neural Networks.

[4]  Brahim Chaib-draa,et al.  Parallel Rollout for Online Solution of Dec-POMDPs , 2008, FLAIRS.

[5]  Alain Dutech,et al.  Using linear programming duality for solving finite horizon Dec-POMDPs , 2008 .

[6]  Manuela M. Veloso,et al.  Exploiting factored representations for decentralized execution in multiagent teams , 2007, AAMAS '07.

[7]  Robert G. Reynolds,et al.  Cultural algorithms: modeling of how cultures learn to solve problems , 2004, 16th IEEE International Conference on Tools with Artificial Intelligence.

[8]  Julita Vassileva,et al.  An influence diagram model for multi-agent negotiation , 2000, Proceedings Fourth International Conference on MultiAgent Systems.

[9]  Shlomo Zilberstein,et al.  Achieving goals in decentralized POMDPs , 2009, AAMAS.

[10]  Edmund H. Durfee,et al.  Recursive Agent Modeling Using Limited Rationality , 1995, ICMAS.

[11]  Yang Xiang,et al.  PROBABILISTIC REASONING IN MULTIAGENT SYSTEMS: A GRAPHICAL MODELS APPROACH, by Yang Xiang, Cambridge University Press, Cambridge, 2002, xii + 294 pp., ISBN 0-521-81308-5 (Hardback, £45.00). , 2002, Robotica.

[12]  Thomas Stützle,et al.  Ant colony optimization: artificial ants as a computational intelligence technique , 2006 .

[13]  Yifeng Zeng,et al.  Graphical models for interactive POMDPs: representations and solutions , 2009, Autonomous Agents and Multi-Agent Systems.

[14]  Yang Xiang,et al.  Comparison of tightly and loosely coupled decision paradigms in multiagent expedition , 2010, Int. J. Approx. Reason..

[15]  Wolfram Burgard,et al.  Probabilistic Robotics (Intelligent Robotics and Autonomous Agents) , 2005 .

[16]  Sanguk Noh,et al.  Coordination and belief update in a distributed anti-air environment , 1998, Proceedings of the Thirty-First Hawaii International Conference on System Sciences.

[17]  Charles E. Taylor Adaptation in Natural and Artificial Systems: An Introductory Analysis with Applications to Biology, Control, and Artificial Intelligence. Complex Adaptive Systems.John H. Holland , 1994 .

[18]  Shlomo Zilberstein,et al.  Improved Memory-Bounded Dynamic Programming for Decentralized POMDPs , 2007, UAI.

[19]  Neil Immerman,et al.  The Complexity of Decentralized Control of Markov Decision Processes , 2000, UAI.

[20]  Claudia V. Goldman,et al.  Solving Transition Independent Decentralized Markov Decision Processes , 2004, J. Artif. Intell. Res..

[21]  H. Raiffa,et al.  Decisions with Multiple Objectives , 1993 .

[22]  David E. Smith,et al.  Planning Under Continuous Time and Resource Uncertainty: A Challenge for AI , 2002, AIPS Workshop on Planning for Temporal Domains.

[23]  Hiroaki Kitano,et al.  RoboCup Rescue: search and rescue in large-scale disasters as a domain for autonomous agents research , 1999, IEEE SMC'99 Conference Proceedings. 1999 IEEE International Conference on Systems, Man, and Cybernetics (Cat. No.99CH37028).

[24]  Makoto Yokoo,et al.  Taming Decentralized POMDPs: Towards Efficient Policy Computation for Multiagent Settings , 2003, IJCAI.

[25]  Martha E. Pollack,et al.  Introducing the Tileworld: Experimentally Evaluating Agent Architectures , 1990, AAAI.

[26]  Leslie Pack Kaelbling,et al.  Learning Policies for Partially Observable Environments: Scaling Up , 1997, ICML.

[27]  Peter Norvig,et al.  Artificial Intelligence: A Modern Approach , 1995 .

[28]  Shimon Whiteson,et al.  Exploiting locality of interaction in factored Dec-POMDPs , 2008, AAMAS.

[29]  Yang Xiang Tractable Optimal Multiagent Collaborative Design , 2007, IAT.

[30]  Bernard Manderick,et al.  Modeling a Multi-Agent Environment Combining Influence Diagrams , 2007 .

[31]  Craig Boutilier,et al.  Decision-Theoretic Planning: Structural Assumptions and Computational Leverage , 1999, J. Artif. Intell. Res..

[32]  William S. Havens,et al.  Optimal design in collaborative design network , 2005, AAMAS '05.