暂无分享,去创建一个
Tom Eccles | Rahul Savani | Yoram Bachrach | Andrea Tacchetti | Marc Lanctot | Ian Gemp | Richard Everett | Thomas Anthony | J'anos Kram'ar
[1] Wojciech M. Czarnecki,et al. Grandmaster level in StarCraft II using multi-agent reinforcement learning , 2019, Nature.
[2] Sriram Srinivasan,et al. OpenSpiel: A Framework for Reinforcement Learning in Games , 2019, ArXiv.
[3] Xiaotie Deng,et al. Settling the Complexity of Two-Player Nash Equilibrium , 2006, 2006 47th Annual IEEE Symposium on Foundations of Computer Science (FOCS'06).
[4] Kevin Waugh,et al. DeepStack: Expert-level artificial intelligence in heads-up no-limit poker , 2017, Science.
[5] John Fearnley,et al. An Empirical Study on Computing Equilibria in Polymatrix Games , 2016, AAMAS.
[6] H. Kuk. On equilibrium points in bimatrix games , 1996 .
[7] Sebastian Nowozin,et al. The Numerics of GANs , 2017, NIPS.
[8] Adam Lerer,et al. Combining Deep Reinforcement Learning and Search for Imperfect-Information Games , 2020, NeurIPS.
[9] Kousha Etessami,et al. On the Complexity of Nash Equilibria and Other Fixed Points , 2010, SIAM J. Comput..
[10] Xiaotie Deng,et al. Settling the complexity of computing two-player Nash equilibria , 2007, JACM.
[11] Paul G. Spirakis,et al. Computing Approximate Nash Equilibria in Polymatrix Games , 2015, Algorithmica.
[12] R. McKelvey,et al. Computation of equilibria in finite games , 1996 .
[13] Paul W. Goldberg,et al. The complexity of computing a Nash equilibrium , 2006, STOC '06.
[14] Philip B. Stark,et al. From Battlefields to Elections: Winning Strategies of Blotto and Auditing Games , 2018, SODA.
[15] Karl Tuyls,et al. Computing Approximate Equilibria in Sequential Adversarial Games by Exploitability Descent , 2019, IJCAI.
[16] Mohammad Taghi Hajiaghayi,et al. Faster and Simpler Algorithm for Optimal Strategies of Blotto Game , 2016, AAAI.
[17] Daphne Koller,et al. A Continuation Method for Nash Equilibria in Structured Games , 2003, IJCAI.
[18] Michael P. Wellman,et al. Empirical game-theoretic analysis of the TAC Supply Chain game , 2007, AAMAS '07.
[19] D. Whitehead,et al. The El Farol Bar Problem Revisited: Reinforcement Learning in a Potential Game , 2008 .
[20] A. Juditsky,et al. Solving variational inequalities with Stochastic Mirror-Prox algorithm , 2008, 0809.0815.
[21] Noam Brown,et al. Superhuman AI for heads-up no-limit poker: Libratus beats top professionals , 2018, Science.
[22] Robert Wilson,et al. A global Newton method to compute Nash equilibria , 2003, J. Econ. Theory.
[23] Mohammad Taghi Hajiaghayi,et al. From Duels to Battlefields: Computing Equilibria of Blotto and Other Games , 2016, AAAI.
[24] F. Facchinei,et al. Finite-Dimensional Variational Inequalities and Complementarity Problems , 2003 .
[25] Alain Durmus,et al. Convergence Analysis of Riemannian Stochastic Approximation Schemes , 2020, ArXiv.
[26] Yoav Shoham,et al. Multiagent Systems - Algorithmic, Game-Theoretic, and Logical Foundations , 2009 .
[27] Jie Chen,et al. Stochastic Gradient Descent with Biased but Consistent Gradient Estimators , 2018, ArXiv.
[28] Georg Ostrovski,et al. Payoff Performance of Fictitious Play , 2013, ArXiv.
[29] Neil Burch,et al. Heads-up limit hold’em poker is solved , 2015, Science.
[30] Robert Wilson,et al. Computing Nash equilibria by iterated polymatrix approximation , 2004 .
[31] Jonathan Gray,et al. Human-Level Performance in No-Press Diplomacy via Equilibrium Search , 2020, ICLR.
[32] D. Blackwell. An analog of the minimax theorem for vector payoffs. , 1956 .
[33] David M. Pennock,et al. An Empirical Game-Theoretic Analysis of Price Discovery in Prediction Markets , 2016, IJCAI.
[34] Ayala Arad,et al. Multi-dimensional iterative reasoning in action: The case of the Colonel Blotto game , 2012 .
[35] Marc Teboulle,et al. Mirror descent and nonlinear projected subgradient methods for convex optimization , 2003, Oper. Res. Lett..
[36] Avrim Blum,et al. Planning in the Presence of Cost Functions Controlled by an Adversary , 2003, ICML.
[37] Cao Xiao,et al. FastGCN: Fast Learning with Graph Convolutional Networks via Importance Sampling , 2018, ICLR.
[38] Troels Bjerre Lund,et al. On the approximation performance of fictitious play in finite games , 2011, Int. J. Game Theory.
[39] W. Arthur. Complexity in economic theory: inductive reasoning and bounded rationality , 1994 .
[40] Georgios Piliouras,et al. No-regret learning and mixed Nash equilibria: They do not mix , 2020, NeurIPS.
[41] David Silver,et al. A Unified Game-Theoretic Approach to Multiagent Reinforcement Learning , 2017, NIPS.
[42] Michael P. Wellman. Methods for Empirical Game-Theoretic Analysis , 2006, AAAI.
[43] G. M. Korpelevich. The extragradient method for finding saddle points and other problems , 1976 .
[44] Christos H. Papadimitriou,et al. Cycles in adversarial regularized learning , 2017, SODA.
[45] Yoav Shoham,et al. Simple search methods for finding a Nash equilibrium , 2004, Games Econ. Behav..
[46] Mohammad Taghi Hajiaghayi,et al. Optimal Strategies of Blotto Games: Beyond Convexity , 2019, EC.
[47] Theodore L. Turocy. A dynamic homotopy interpretation of the logistic quantal response equilibrium correspondence , 2005, Games Econ. Behav..
[48] R. McKelvey,et al. Quantal Response Equilibria for Normal Form Games , 1995 .
[49] Enric Boix-Adsera,et al. The Multiplayer Colonel Blotto Game , 2020, EC.
[50] Yoav Shoham,et al. Run the GAMUT: a comprehensive approach to evaluating game-theoretic algorithms , 2004, Proceedings of the Third International Joint Conference on Autonomous Agents and Multiagent Systems, 2004. AAMAS 2004..
[51] John Fearnley,et al. Finding Approximate Nash Equilibria of Bimatrix Games via Payoff Queries , 2013, ACM Trans. Economics and Comput..
[52] Tom Eccles,et al. Learning to Play No-Press Diplomacy with Best Response Policy Iteration , 2020, NeurIPS.
[53] Yakov Babichenko,et al. Query complexity of approximate nash equilibria , 2013, STOC.
[54] A. Talman,et al. Simplicial variable dimension algorithms for solving the nonlinear complementarity problem on a product of unit simplices using a general labelling , 1987 .
[55] Mark Fey,et al. Symmetric games with only asymmetric equilibria , 2012, Games Econ. Behav..
[56] Paul W. Goldberg,et al. Learning equilibria of games via payoff queries , 2013, EC '13.
[57] Andrew McLennan,et al. Gambit: Software Tools for Game Theory , 2006 .
[58] R. Sutton,et al. A convergent O ( n ) algorithm for off-policy temporal-difference learning with linear function approximation , 2008, NIPS 2008.
[59] Vincent Conitzer,et al. Mixed-Integer Programming Methods for Finding Nash Equilibria , 2005, AAAI.
[60] D. Fudenberg,et al. The Theory of Learning in Games , 1998 .
[61] Georgios Piliouras,et al. From Poincaré Recurrence to Convergence in Imperfect Information Games: Finding Equilibrium via Regularization , 2020, ICML.