Performance Models for Large Scale Multi-Agent System: A Distributed Pomdp-Based Approach

Given a large group of cooperative agents, selecting the right coordination or conflict resolution strategy can have a significant impact on their performance (e.g., speed of convergence). While performance models of such coordination or conflict resolution strategies could aid in selecting the right strategy for a given domain, such models remain largely uninvestigated in the multiagent literature. This chapter takes a step towards applying the recently emerging distributed POMDP (partially observable Markov decision process) frameworks, such as MTDP (Markov team decision process), in service of creating such performance models. A strategy is mapped onto an MTDP policy, and strategies are compared by evaluating their corresponding policies. To address issues of scale-up in applying the distributed POMDP-based models, we use small-scale models, called building blocks that represent the local interaction among a small group of agents. We discuss several ways to combine building blocks for performance prediction of a larger-scale multiagent system.

[1]  Leslie Pack Kaelbling,et al.  On the Complexity of Solving Markov Decision Problems , 1995, UAI.

[2]  Victor R. Lesser,et al.  Multi-agent policies: from centralized ones to decentralized ones , 2002, AAMAS '02.

[3]  Makoto Yokoo,et al.  Distributed constraint satisfaction algorithm for complex local problems , 1998, Proceedings International Conference on Multi Agent Systems (Cat. No.98EX160).

[4]  Marius-Calin Silaghi,et al.  Asynchronous Search with Aggregations , 2000, AAAI/IAAI.

[5]  Martin L. Puterman,et al.  Markov Decision Processes: Discrete Stochastic Dynamic Programming , 1994 .

[6]  Ronald Parr,et al.  Flexible Decomposition Algorithms for Weakly Coupled Markov Decision Problems , 1998, UAI.

[7]  Steven Minton,et al.  Solving Large-Scale Constraint-Satisfaction and Scheduling Problems Using a Heuristic Repair Method , 1990, AAAI.

[8]  Thomas Dean,et al.  Decomposition Techniques for Planning in Stochastic Domains , 1995, IJCAI.

[9]  Joelle Pineau,et al.  A Hierarchical Approach to POMDP Planning and Execution , 2004 .

[10]  Bart Selman,et al.  Heavy-Tailed Phenomena in Satisfiability and Constraint Satisfaction Problems , 2000, Journal of Automated Reasoning.

[11]  Nicholas R. Jennings,et al.  Learning to select a coordination mechanism , 2002, AAMAS '02.

[12]  Neil Immerman,et al.  The Complexity of Decentralized Control of Markov Decision Processes , 2000, UAI.

[13]  Milind Tambe,et al.  Argumentation as distributed constraint satisfaction: applications and results , 2001, AGENTS '01.

[14]  Boi Faltings,et al.  Secure Asynchronous Search , 2001 .

[15]  Michel Lemaître,et al.  Branch and Bound Algorithm Selection by Performance Prediction , 1998, AAAI/IAAI.

[16]  Milind Tambe,et al.  The Communicative Multiagent Team Decision Problem: Analyzing Teamwork Theories and Models , 2011, J. Artif. Intell. Res..

[17]  Stuart J. Russell,et al.  How Long Will It Take? , 1992, AAAI.

[18]  Weixiong Zhang,et al.  Distributed breakout revisited , 2002, AAAI/IAAI.

[19]  Omer F. Rana,et al.  Performance management of mobile agent systems , 2000, AGENTS '00.

[20]  Wei-Min Shen,et al.  A Dynamic Distributed Constraint Satisfaction Approach to Resource Allocation , 2001, CP.

[21]  Victor R. Lesser,et al.  The Use of Meta-level Information in Learning Situation-Specific Coordination , 1997, IJCAI.