Hindsight and Sequential Rationality of Correlated Play
暂无分享,去创建一个
Michael Bowling | Amy Greenwald | Marc Lanctot | Dustin Morrill | Reca Sarfati | Ryan D'Orazio | James Wright
[1] S. Hart,et al. A simple adaptive procedure leading to correlated equilibrium , 2000 .
[2] J. Vial,et al. Strategically zero-sum games: The class of games whose completely mixed equilibria cannot be improved upon , 1978 .
[3] S. Vajda. Some topics in two-person games , 1971 .
[4] Amy Greenwald,et al. A General Class of No-Regret Learning Algorithms and Game-Theoretic Equilibria , 2003, COLT.
[5] H. W. Kuhn,et al. 11. Extensive Games and the Problem of Information , 1953 .
[6] Tuomas Sandholm,et al. Coarse Correlation in Extensive-Form Games , 2019, AAAI.
[7] Wojciech M. Czarnecki,et al. Grandmaster level in StarCraft II using multi-agent reinforcement learning , 2019, Nature.
[8] Dean P. Foster,et al. Calibrated Learning and Correlated Equilibrium , 1997 .
[9] Miroslav Dudík,et al. A Sampling-Based Approach to Computing Equilibria in Succinct Extensive-Form Games , 2009, UAI.
[10] Demis Hassabis,et al. A general reinforcement learning algorithm that masters chess, shogi, and Go through self-play , 2018, Science.
[11] Demis Hassabis,et al. Mastering the game of Go without human knowledge , 2017, Nature.
[12] Demis Hassabis,et al. Mastering Chess and Shogi by Self-Play with a General Reinforcement Learning Algorithm , 2017, ArXiv.
[13] Martin Schmid,et al. Revisiting CFR+ and Alternating Updates , 2018, J. Artif. Intell. Res..
[14] Ronald J. Williams,et al. Simple Statistical Gradient-Following Algorithms for Connectionist Reinforcement Learning , 2004, Machine Learning.
[15] Yishay Mansour,et al. Policy Gradient Methods for Reinforcement Learning with Function Approximation , 1999, NIPS.
[16] Michael H. Bowling,et al. Actor-Critic Policy Optimization in Partially Observable Multiagent Environments , 2018, NeurIPS.
[17] R. Aumann. Subjectivity and Correlation in Randomized Strategies , 1974 .
[18] F. Forges,et al. Computionally Efficient Coordination in Games Trees , 2002 .
[19] Nicola Gatti,et al. Learning to Correlate in Multi-Player General-Sum Sequential Games , 2019, NeurIPS.
[20] Eric van Damme,et al. Non-Cooperative Games , 2000 .
[21] Dean P. Foster,et al. Regret in the On-Line Decision Problem , 1999 .
[22] Kevin Waugh,et al. DeepStack: Expert-level artificial intelligence in heads-up no-limit poker , 2017, Science.
[23] Noam Brown,et al. Superhuman AI for multiplayer poker , 2019, Science.
[24] R. Selten. Reexamination of the perfectness concept for equilibrium points in extensive games , 1975, Classics in Game Theory.
[25] Sriram Srinivasan,et al. OpenSpiel: A Framework for Reinforcement Learning in Games , 2019, ArXiv.
[26] Bernhard von Stengel,et al. Extensive-Form Correlated Equilibrium: Definition and Computational Complexity , 2008, Math. Oper. Res..
[27] Demis Hassabis,et al. Mastering the game of Go with deep neural networks and tree search , 2016, Nature.
[28] James Hannan,et al. 4. APPROXIMATION TO RAYES RISK IN REPEATED PLAY , 1958 .
[29] Michael H. Bowling,et al. Regret Minimization in Games with Incomplete Information , 2007, NIPS.
[30] Noam Brown,et al. Superhuman AI for heads-up no-limit poker: Libratus beats top professionals , 2018, Science.