暂无分享,去创建一个
Sam Devlin | Spyridon Samothrakis | James Alfred Walker | Daniel Hernandez | Kevin Denamganaï | Spyridon Samothrakis | Sam Devlin | Kevin Denamganaï | Daniel Hernández
[1] Matthew E. Taylor,et al. A survey and critique of multiagent deep reinforcement learning , 2018, Autonomous Agents and Multi-Agent Systems.
[2] Alec Radford,et al. Proximal Policy Optimization Algorithms , 2017, ArXiv.
[3] David Silver,et al. A Unified Game-Theoretic Approach to Multiagent Reinforcement Learning , 2017, NIPS.
[4] Thore Graepel,et al. Re-evaluating evaluation , 2018, NeurIPS.
[5] Arthur L. Samuel,et al. Some Studies in Machine Learning Using the Game of Checkers , 1967, IBM J. Res. Dev..
[6] Gerald Tesauro,et al. Neurogammon: a neural-network backgammon program , 1990, 1990 IJCNN International Joint Conference on Neural Networks.
[7] Marlos C. Machado,et al. Domain-Independent Optimistic Initialization for Reinforcement Learning , 2014, AAAI Workshop: Learning for General Competency in Video Games.
[8] Tom Schaul,et al. Prioritized Experience Replay , 2015, ICLR.
[9] Jakub W. Pachocki,et al. Emergent Complexity via Multi-Agent Competition , 2017, ICLR.
[10] Peter L. Bartlett,et al. RL$^2$: Fast Reinforcement Learning via Slow Reinforcement Learning , 2016, ArXiv.
[11] N. Le Fort-Piat,et al. The world of independent learners is not markovian , 2011, Int. J. Knowl. Based Intell. Eng. Syst..
[12] Carlos Cotta,et al. An Analysis of Hall-of-Fame Strategies in Competitive Coevolutionary Algorithms for Self-Learning in RTS Games , 2013, LION.
[13] Gerald Tesauro,et al. Temporal Difference Learning and TD-Gammon , 1995, J. Int. Comput. Games Assoc..
[14] Geoffrey E. Hinton,et al. Visualizing Data using t-SNE , 2008 .
[15] Pieter Abbeel,et al. Continuous Adaptation via Meta-Learning in Nonstationary and Competitive Environments , 2017, ICLR.
[16] David Barber,et al. Thinking Fast and Slow with Deep Learning and Tree Search , 2017, NIPS.
[17] Marco Wiering,et al. Reinforcement learning in the game of Othello: Learning against a fixed opponent and learning from self-play , 2013, 2013 IEEE Symposium on Adaptive Dynamic Programming and Reinforcement Learning (ADPRL).
[18] Joel Z. Leibo,et al. Human-level performance in first-person multiplayer games with population-based deep reinforcement learning , 2018, ArXiv.
[19] Karl Tuyls,et al. alpha-Rank: Multi-Agent Evaluation by Evolution , 2019 .
[20] Sam Devlin,et al. A Generalized Framework for Self-Play Training , 2019, 2019 IEEE Conference on Games (CoG).
[21] Guy Lever,et al. Human-level performance in 3D multiplayer games with population-based reinforcement learning , 2018, Science.
[22] Wojciech M. Czarnecki,et al. Grandmaster level in StarCraft II using multi-agent reinforcement learning , 2019, Nature.
[23] Guy Lever,et al. A Generalized Training Approach for Multiagent Learning , 2020, ICLR.
[24] Demis Hassabis,et al. Mastering the game of Go without human knowledge , 2017, Nature.
[25] Demis Hassabis,et al. Mastering Chess and Shogi by Self-Play with a General Reinforcement Learning Algorithm , 2017, ArXiv.
[26] Ronald J. Williams,et al. Simple Statistical Gradient-Following Algorithms for Connectionist Reinforcement Learning , 2004, Machine Learning.
[27] Max Jaderberg,et al. Open-ended Learning in Symmetric Zero-sum Games , 2019, ICML.
[28] Michal Valko,et al. Multiagent Evaluation under Incomplete Information , 2019, NeurIPS.