Approximation guarantees for fictitious play

Fictitious play is a simple, well-known, and often-used algorithm for playing (and, especially, learning to play) games. However, in general it does not converge to equilibrium; even when it does, we may not be able to run it to convergence. Still, we may obtain an approximate equilibrium. In this paper, we study the approximation properties that fictitious play obtains when it is run for a limited number of rounds. We show that if both players randomize uniformly over their actions in the first r rounds of fictitious play, then the result is an e-equilibrium, where ∊ = (r + l)/(2r). (Since we are examining only a constant number of pure strategies, we know that ∊ ≤ 1/2 is impossible, due to a result of Feder et al.) We show that this bound is tight in the worst case; however, with an experiment on random games, we illustrate that fictitious play usually obtains a much better approximation. We then consider the possibility that the players fail to choose the same r. We show how to obtain the optimal approximation guarantee when both the opponent's r and the game are adversarially chosen (but there is an upper bound R on the opponent's r), using a linear program formulation. We show that if the action played in the ith round of fictitious play is chosen with probability proportional to: 1 for i = 1 and l/(i − 1) for all 2 ≤ i ≤ R + l, this gives an approximation guarantee of 1 − 1/(2 + lnÄ). We also obtain a lower bound of 1 − 4/ In R. This provides an actionable prescription for how long to run fictitious play.

[1]  J. Robinson AN ITERATIVE METHOD OF SOLVING A GAME , 1951, Classics in Game Theory.

[2]  O. H. Brownlee,et al.  ACTIVITY ANALYSIS OF PRODUCTION AND ALLOCATION , 1952 .

[3]  K. Miyasawa ON THE CONVERGENCE OF THE LEARNING PROCESS IN A 2 X 2 NON-ZERO-SUM TWO-PERSON GAME , 1961 .

[4]  C. E. Lemke,et al.  Equilibrium Points of Bimatrix Games , 1964 .

[5]  L. Shapley SOME TOPICS IN TWO-PERSON GAMES , 1963 .

[6]  Eitan Zemel,et al.  Nash and correlated equilibria: Some complexity considerations , 1989 .

[7]  John Nachbar “Evolutionary” selection dynamics in games: Convergence and limit properties , 1990 .

[8]  I. Althöfer On sparse approximations to randomized strategies and convex combinations , 1994 .

[9]  L. Shapley,et al.  Fictitious Play Property for Games with Identical Interests , 1996 .

[10]  Michael L. Littman,et al.  Abstraction Methods for Game Theoretic Poker , 2000, Computers and Games.

[11]  Manuela M. Veloso,et al.  Multiagent learning using a variable learning rate , 2002, Artif. Intell..

[12]  Vincent Conitzer,et al.  AWESOME: A general multiagent learning algorithm that converges in self-play and learns a best response against stationary opponents , 2003, Machine Learning.

[13]  Aranyak Mehta,et al.  Playing large games using simple strategies , 2003, EC '03.

[14]  Michael H. Bowling,et al.  Convergence and No-Regret in Multiagent Learning , 2004, NIPS.

[15]  Xi Chen,et al.  3-NASH is PPAD-Complete , 2005, Electron. Colloquium Comput. Complex..

[16]  Vincent Conitzer,et al.  Mixed-Integer Programming Methods for Finding Nash Equilibria , 2005, AAAI.

[17]  Tuomas Sandholm,et al.  Optimal Rhode Island Hold'em Poker , 2005, AAAI.

[18]  Tuomas Sandholm,et al.  Optimal Rhode Island Hold'em Poker , 2005, AAAI.

[19]  Christos H. Papadimitriou,et al.  Three-Player Games Are Hard , 2005, Electron. Colloquium Comput. Complex..

[20]  Xiaotie Deng,et al.  Settling the Complexity of Two-Player Nash Equilibrium , 2006, 2006 47th Annual IEEE Symposium on Foundations of Computer Science (FOCS'06).

[21]  Rahul Savani,et al.  Hard‐to‐Solve Bimatrix Games , 2006 .

[22]  William Dudziak Using Fictitious Play to Find Pseudo-optimal Solutions for Full-scale Poker , 2006, IC-AI.

[23]  Aranyak Mehta,et al.  A Note on Approximate Nash Equilibria , 2006, WINE.

[24]  Vincent Conitzer,et al.  AWESOME: A general multiagent learning algorithm that converges in self-play and learns a best response against stationary opponents , 2003, Machine Learning.

[25]  Bikramjit Banerjee,et al.  RVσ(t): a unifying approach to performance and convergence in online multiagent learning , 2006, AAMAS '06.

[26]  Bikramjit Banerjee,et al.  RVσ(t): a unifying approach to performance and convergence in online multiagent learning , 2006, AAMAS '06.

[27]  Amin Saberi,et al.  Approximating nash equilibria using small-support strategies , 2007, EC '07.

[28]  Yoav Shoham,et al.  If multi-agent learning is the answer, what is the question? , 2007, Artif. Intell..

[29]  Paul G. Spirakis,et al.  Efficient Algorithms for Constant Well Supported Approximate Equilibria in Bimatrix Games , 2007, ICALP.

[30]  Paul G. Spirakis,et al.  An Optimization Approach for Approximate Nash Equilibria , 2007, WINE.

[31]  Aranyak Mehta,et al.  Progress in approximate nash equilibria , 2007, EC '07.

[32]  Tuomas Sandholm,et al.  Non-commercial Research and Educational Use including without Limitation Use in Instruction at Your Institution, Sending It to Specific Colleagues That You Know, and Providing a Copy to Your Institution's Administrator. All Other Uses, Reproduction and Distribution, including without Limitation Comm , 2022 .

[33]  Geoffrey J. Gordon,et al.  A Fast Bundle-based Anytime Algorithm for Poker and other Convex Games , 2007, AISTATS.

[34]  Yoav Shoham,et al.  Simple search methods for finding a Nash equilibrium , 2004, Games Econ. Behav..

[35]  Paul G. Spirakis,et al.  Approximate Equilibria for Strategic Two Person Games , 2008, SAGT.

[36]  Tuomas Sandholm,et al.  Computing an approximate jam/fold equilibrium for 3-player no-limit Texas Hold'em tournaments , 2008, AAMAS.

[37]  Vincent Conitzer,et al.  New complexity results about Nash equilibria , 2008, Games Econ. Behav..

[38]  Vincent Conitzer,et al.  New complexity results about Nash equilibria , 2008, Games Econ. Behav..

[39]  Paul W. Goldberg,et al.  The Complexity of Computing a Nash Equilibrium , 2009, SIAM J. Comput..

[40]  Evangelos Markakis,et al.  New algorithms for approximate Nash equilibria in bimatrix games , 2010, Theor. Comput. Sci..