Implicit Negotiation in Repeated Games

In business-related interactions such as the on-going high-stakes FCC spectrum auctions, explicit communication among participants is regarded as collusion, and is therefore illegal. In this paper, we consider the possibility of autonomous agents engaging in implicit negotiation via their tacit interactions. In repeated general-sum games, our testbed for studying this type of interaction, an agent using a "best response" strategy maximizes its own payoff assuming its behavior has no effect on its opponent. This notion of best response requires some degree of learning to determine the fixed opponent behavior. Against an unchanging opponent, the best-response agent performs optimally, and can be thought of as a "follower," since it adapts to its opponent. However, pairing two best-response agents in a repeated game can result in suboptimal behavior. We demonstrate this suboptimality in several different games using variants of Q-learning as an example of a best-response strategy.We then examine two "leader" strategies that induce better performance from opponent followers via stubbornness and threats. These tactics are forms of implicit negotiation in that they aim to achieve a mutually beneficial outcome without using explicit communication outside of the game.

[1]  E. Rowland Theory of Games and Economic Behavior , 1946, Nature.

[2]  Sarit Kraus,et al.  Coordination without Communication: Experimental Validation of Focal Point Techniques , 1997, ICMAS.

[3]  Jesse A. Schwartz,et al.  Collusive Bidding: Lessons from the FCC Spectrum Auctions , 2000 .

[4]  Peter Stone,et al.  Randomized strategic demand reduction: getting more by asking for less , 2002, AAMAS '02.

[5]  C. Papadimitriou On platers with a bounded number of states , 1992 .

[6]  Eric van Damme,et al.  Non-Cooperative Games , 2000 .

[7]  Peter Stone,et al.  FAucS : An FCC Spectrum Auction Simulator for Autonomous Bidding Agents , 2001, WELCOM.

[8]  Craig Boutilier,et al.  The Dynamics of Reinforcement Learning in Cooperative Multiagent Systems , 1998, AAAI/IAAI.

[9]  Jeffrey S. Rosenschein,et al.  Mechanisms for Automated Negotiation in State Oriented Domains , 1996, J. Artif. Intell. Res..

[10]  Gunes Ercal,et al.  On No-Regret Learning, Fictitious Play, and Nash Equilibrium , 2001, ICML.

[11]  D. Fudenberg,et al.  The Theory of Learning in Games , 1998 .

[12]  Michael P. Wellman,et al.  Multiagent Reinforcement Learning: Theoretical Framework and an Algorithm , 1998, ICML.

[13]  Ariel Rubinstein,et al.  A Course in Game Theory , 1995 .

[14]  Michael L. Littman,et al.  Friend-or-Foe Q-learning in General-Sum Games , 2001, ICML.

[15]  R. Weber Making More from Less: Strategic Demand Reduction in the FCC Spectrum Auctions , 1997 .

[16]  Sandip Sen,et al.  Evaluating concurrent reinforcement learners , 2000, Proceedings Fourth International Conference on MultiAgent Systems.

[17]  Victor R. Lesser,et al.  Issues in Automated Negotiation and Electronic Commerce: Extending the Contract Net Framework , 1997, ICMAS.

[18]  Manuela M. Veloso,et al.  Task Decomposition, Dynamic Role Assignment, and Low-Bandwidth Communication for Real-Time Strategic Teamwork , 1999, Artif. Intell..

[19]  Scott Shenker,et al.  Learning in Network Contexts: Experimental Results from Simulations , 2001, Games Econ. Behav..

[20]  W. Hamilton,et al.  The evolution of cooperation. , 1984, Science.