Adaptive multi-robot coordination: A game-theoretic perspective

Multi-robot systems researchers have been investigating adaptive coordination methods for improving spatial coordination in teams. Such methods adapt the coordination method to the dynamic changes in density of the robots. Unfortunately, while their empirical success is evident, none of these methods has been understood in the context of existing formal work on multi-robot learning. This paper presents a reinforcement-learning approach to coordination algorithm selection, which is not only shown to work well in experiments, but is also analytically grounded. We present a reward function (Effectiveness Index, EI), that reduces time and resources spent coordinating, and maximizes the time between conflicts that require coordination. It does this by measuring the resource-spending velocity. We empirically show its success in simulations of multi-robot foraging. In addition, we analytically explore the reasons that EI works well. We show that under some assumptions, spatial coordination opportunities can be modeled as matrix games in which the payoffs are directly a function of EI estimates. The use of reinforcement learning leads to robots maximizing their EI rewards in equilibrium. This work is a step towards bridging the gap between the theoretical study of interactions, and their use in multi-robot coordination.

[1]  Richard T. Vaughan,et al.  Reducing spatial interference in robot teams by local-investment aggression , 2005, 2005 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[2]  Karl Tuyls,et al.  An Overview of Cooperative and Competitive Multiagent Learning , 2005, LAMAS.

[3]  Kagan Tumer,et al.  Collective Intelligence, Data Routing and Braess' Paradox , 2002, J. Artif. Intell. Res..

[4]  Maja J. Mataric,et al.  Territorial multi-robot task division , 1998, IEEE Trans. Robotics Autom..

[5]  Peter Secretan Learning , 1965, Mental Health.

[6]  Gaurav S. Sukhatme,et al.  Go ahead, make my day: Robot conflict resolution by aggressive competition , 2000 .

[7]  Dani Goldberg Design and Evaluation of Robust Behavior-Based Controllers for Distributed Multi-Robot Collection Tasks , 2001 .

[8]  Maria Gini,et al.  Performance Evaluation of Multiple Robots in a Search and Retrieval Task , 1998 .

[9]  Glenn Ellison Cooperation in the Prisoner's Dilemma with Anonymous Random Matching , 1994 .

[10]  Lynne E. Parker,et al.  Robot Teams: From Diversity to Polymorphism , 2002 .

[11]  Gaurav S. Sukhatme,et al.  Emergent bucket brigading: a simple mechanisms for improving performance in multi-robot constrained-space foraging tasks , 2001, AGENTS '01.

[12]  A. Pentland,et al.  Collective intelligence , 2006, IEEE Comput. Intell. Mag..

[13]  John V. Howard,et al.  Cooperation in the Prisoner's Dilemma , 1988 .

[14]  Nicholas R. Jennings,et al.  The Dynamic Selection of Coordination Mechanisms , 2004, Autonomous Agents and Multi-Agent Systems.

[15]  Maja J. Mataric,et al.  Reinforcement Learning in the Multi-Robot Domain , 1997, Auton. Robots.

[16]  W. Marsden I and J , 2012 .

[17]  E. Hopkins Learning, Matching and Aggregation , 1995 .

[18]  Sarit Kraus,et al.  A study of mechanisms for improving robotic group performance , 2008, Artif. Intell..

[19]  Pattie Maes,et al.  A Study of Territoriality: The Role of Critical Mass in Adaptive Task Division , 1996 .

[20]  Erfu Yang,et al.  Multiagent Reinforcement Learning for Multi-Robot Systems: A Survey , 2004 .

[21]  Tucker R. Balch Integrating Learning with Motor Schema-Based Control for a Robot Soccer Team , 1997, RoboCup.

[22]  F. Vega-Redondo,et al.  Efficient Equilibrium Selection in Evolutionary Games with Random Matching , 1996 .

[23]  Jay R. Galbraith Designing Complex Organizations , 1973 .

[24]  Kagan Tumer,et al.  General principles of learning-based multi-agent systems , 1999, AGENTS '99.

[25]  Bernhard Nebel,et al.  Dynamic decentralized area partitioning for cooperating cleaning robots , 2002, Proceedings 2002 IEEE International Conference on Robotics and Automation (Cat. No.02CH37292).

[26]  Kagan Tumer,et al.  Analyzing and visualizing multiagent rewards in dynamic and stochastic domains , 2008, Autonomous Agents and Multi-Agent Systems.

[27]  Manuela Veloso,et al.  An Analysis of Stochastic Game Theory for Multiagent Reinforcement Learning , 2000 .