A decentralized approach for convention emergence in multi-agent systems

The field of convention emergence studies how agents involved in repeated coordination games can reach consensus through only local interactions. The literature on this topic is vast and is motivated by human societies, mainly addressing coordination problems between human agents, such as who gets to redial after a dropped telephone call. In contrast, real-world engineering problems, such as coordination in wireless sensor networks, involve agents with limited resources and knowledge and thus pose certain restrictions on the complexity of the coordination mechanisms. Due to these restrictions, strategies proposed for human coordination may not be suitable for engineering applications and need to be further explored in the context of real-world application domains. In this article we take the role of designers of large decentralized multi-agent systems. We investigate the factors that speed up the convergence process of agents arranged in different static and dynamic topologies and under different interaction models, typical for engineering applications. We also study coordination problems both under partial observability and in the presence of faults (or noise). The main contributions of this article are that we propose an approach for emergent coordination, motivated by highly constrained devices, such as wireless nodes and swarm bots, in the absence of a central entity and perform extensive theoretical and empirical studies. Our approach is called Win-Stay Lose-probabilistic-Shift, generalizing two well-known strategies in game theory that have been applied in other domains. We demonstrate that our approach performs well in different settings under limited information and imposes minimal system requirements, due to its simplicity. Moreover, our technique outperforms state-of-the-art coordination mechanisms, guarantees full convergence in any topology and has the property that all convention states are absorbing.

[1]  Debraj Ray,et al.  Evolving Aspirations and Cooperation , 1998 .

[2]  Ingrid Moerman,et al.  Network-wide synchronization in Wireless Sensor Networks , 2012, 2012 19th IEEE Symposium on Communications and Vehicular Technology in the Benelux (SCVT).

[3]  H. Young,et al.  The Evolution of Conventions , 1993 .

[4]  Luc Steels,et al.  The synthetic modeling of language origins , 1997 .

[5]  Vincent A. Traag,et al.  Indirect reciprocity through gossiping can lead to cooperative clusters , 2011, 2011 IEEE Symposium on Artificial Life (ALIFE).

[6]  M. Nowak,et al.  A strategy of win-stay, lose-shift that outperforms tit-for-tat in the Prisoner's Dilemma game , 1993, Nature.

[7]  John G. Kemeny,et al.  Finite Markov chains , 1960 .

[8]  Jeffrey Barrett,et al.  The role of forgetting in the evolution and learning of language , 2009, J. Exp. Theor. Artif. Intell..

[9]  Ann Nowé,et al.  Reinforcement Learning for Self-organizing Wake-Up Scheduling in Wireless Sensor Networks , 2011, ICAART.

[10]  Yoav Shoham,et al.  Dispersion games: general definitions and some specific learning results , 2002, AAAI/IAAI.

[11]  Ramon Sangüesa,et al.  Emergence of coordination in scale-free networks , 2003, Web Intell. Agent Syst..

[12]  Sandip Sen,et al.  Social Instruments for Robust Convention Emergence , 2011, IJCAI.

[13]  M. Hauser The Evolution of Communication , 1996 .

[14]  Sandip Sen,et al.  Exploring the Dimensions of Convention Emergence in Multiagent Systems , 2011, Adv. Complex Syst..

[15]  M. Mihaylov Decentralized Coordination in Multi ‐ Agent Systems , 2012 .

[16]  Nicholas R. Jennings,et al.  Decentralised coordination of low-power embedded devices using the max-sum algorithm , 2008, AAMAS.

[17]  Nancy A. Lynch,et al.  Impossibility of distributed consensus with one faulty process , 1983, PODS '83.

[18]  David B. Knoester,et al.  Evolving Virtual Fireflies , 2009, ECAL.

[19]  Harold H. Kelley,et al.  The development of cooperation in the "minimal social situation". , 1962 .

[20]  Sandip Sen,et al.  Norm emergence under constrained interactions in diverse societies , 2008, AAMAS.

[21]  Bhaskar Krishnamachari,et al.  An adaptive energy-efficient and low-latency MAC for data gathering in wireless sensor networks , 2004, 18th International Parallel and Distributed Processing Symposium, 2004. Proceedings..

[22]  James E. Kittock Emergent Conventions and the Structure of Multi--Agent Systems , 1995 .

[23]  F. C. Santos,et al.  Emergence of Cooperation in Adaptive Social Networks with Behavioral Diversity , 2009, ECAL.

[24]  Deborah Estrin,et al.  Medium access control with coordinated adaptive sleeping for wireless sensor networks , 2004, IEEE/ACM Transactions on Networking.

[25]  Arshad Jhumka,et al.  Manipulating convention emergence using influencer agents , 2013, Autonomous Agents and Multi-Agent Systems.

[26]  Jonathan Bendor,et al.  Aspirations, adaptive learning and cooperation in repeated games , 1994 .

[27]  Moshe Tennenholtz,et al.  On the Emergence of Social Conventions: Modeling, Analysis, and Simulations , 1997, Artif. Intell..

[28]  Reka Albert,et al.  Mean-field theory for scale-free random networks , 1999 .

[29]  Bastin Tony Roy Savarimuthu,et al.  Norm emergence in agent societies formed by dynamically changing networks , 2009, Web Intell. Agent Syst..

[30]  Sanjeev Goyal,et al.  Network formation and anti-coordination games , 2004, Int. J. Game Theory.

[31]  David Lewis Convention: A Philosophical Study , 1986 .

[32]  Sandip Sen,et al.  Modeling the Emergence and Convergence of Norms , 2011, IJCAI.

[33]  Sandip Sen,et al.  Emergence of Norms through Social Learning , 2007, IJCAI.

[34]  Moshe Tennenholtz,et al.  Co-Learning and the Evolution of Social Acitivity , 1994 .

[35]  João Balsa,et al.  Force Versus Majority: A Comparison in Convention Emergence Efficiency , 2009, COIN@AAMAS&AAAI.

[36]  Karl Tuyls,et al.  Learning to Reach Agreement in a Continuous Ultimatum Game , 2008, J. Artif. Intell. Res..

[37]  S. Fortunato,et al.  Statistical physics of social dynamics , 2007, 0710.3256.

[38]  Satoru Takahashi,et al.  Anti-Coordination Games and Dynamic stability , 2007, IGTR.

[39]  M. Mobilia Does a single zealot affect an infinite group of voters? , 2003, Physical review letters.

[40]  T. Schelling,et al.  The Strategy of Conflict. , 1961 .

[41]  W. Hamilton,et al.  The evolution of cooperation. , 1984, Science.

[42]  R. Axelrod An Evolutionary Approach to Norms , 1986, American Political Science Review.

[43]  Manuela M. Veloso,et al.  Multiagent learning using a variable learning rate , 2002, Artif. Intell..