暂无分享,去创建一个
Michael Bowling | Amy Greenwald | Marc Lanctot | Dustin Morrill | Reca Sarfati | James Wright | Ryan D'Orazio
[1] Nicola Gatti,et al. Learning to Correlate in Multi-Player General-Sum Sequential Games , 2019, NeurIPS.
[2] Dean P. Foster,et al. Regret in the On-Line Decision Problem , 1999 .
[3] R. Selten. Reexamination of the perfectness concept for equilibrium points in extensive games , 1975, Classics in Game Theory.
[4] Noam Brown,et al. Superhuman AI for multiplayer poker , 2019, Science.
[5] Noam Brown,et al. Superhuman AI for heads-up no-limit poker: Libratus beats top professionals , 2018, Science.
[6] Bernhard von Stengel,et al. Extensive-Form Correlated Equilibrium: Definition and Computational Complexity , 2008, Math. Oper. Res..
[7] R. Vohra,et al. Calibrated Learning and Correlated Equilibrium , 1996 .
[8] R. Aumann. Subjectivity and Correlation in Randomized Strategies , 1974 .
[9] F. Forges,et al. Computionally Efficient Coordination in Games Trees , 2002 .
[10] S. Hart,et al. A simple adaptive procedure leading to correlated equilibrium , 2000 .
[11] Michael H. Bowling,et al. Actor-Critic Policy Optimization in Partially Observable Multiagent Environments , 2018, NeurIPS.
[12] J. Vial,et al. Strategically zero-sum games: The class of games whose completely mixed equilibria cannot be improved upon , 1978 .
[13] Eric van Damme,et al. Non-Cooperative Games , 2000 .
[14] Wojciech M. Czarnecki,et al. Grandmaster level in StarCraft II using multi-agent reinforcement learning , 2019, Nature.
[15] H. W. Kuhn,et al. 11. Extensive Games and the Problem of Information , 1953 .
[16] Demis Hassabis,et al. A general reinforcement learning algorithm that masters chess, shogi, and Go through self-play , 2018, Science.
[17] Yishay Mansour,et al. Policy Gradient Methods for Reinforcement Learning with Function Approximation , 1999, NIPS.
[18] Ronald J. Williams,et al. Simple Statistical Gradient-Following Algorithms for Connectionist Reinforcement Learning , 2004, Machine Learning.
[19] Demis Hassabis,et al. Mastering the game of Go with deep neural networks and tree search , 2016, Nature.
[20] Amy Greenwald,et al. A General Class of No-Regret Learning Algorithms and Game-Theoretic Equilibria , 2003, COLT.
[21] Demis Hassabis,et al. Mastering Chess and Shogi by Self-Play with a General Reinforcement Learning Algorithm , 2017, ArXiv.
[22] Sriram Srinivasan,et al. OpenSpiel: A Framework for Reinforcement Learning in Games , 2019, ArXiv.
[23] Michael H. Bowling,et al. Regret Minimization in Games with Incomplete Information , 2007, NIPS.
[24] S. Vajda. Some topics in two-person games , 1971 .
[25] Demis Hassabis,et al. Mastering the game of Go without human knowledge , 2017, Nature.
[26] James Hannan,et al. 4. APPROXIMATION TO RAYES RISK IN REPEATED PLAY , 1958 .
[27] Martin Schmid,et al. Revisiting CFR+ and Alternating Updates , 2018, J. Artif. Intell. Res..
[28] Kevin Waugh,et al. DeepStack: Expert-level artificial intelligence in heads-up no-limit poker , 2017, Science.
[29] Tuomas Sandholm,et al. Coarse Correlation in Extensive-Form Games , 2019, AAAI.
[30] Alberto Marchesi,et al. No-Regret Learning Dynamics for Extensive-Form Correlated Equilibrium , 2020, NeurIPS.
[31] Miroslav Dudík,et al. A Sampling-Based Approach to Computing Equilibria in Succinct Extensive-Form Games , 2009, UAI.