STARDATA: A StarCraft AI Research Dataset
暂无分享,去创建一个
Gabriel Synnaeve | Zeming Lin | Vasil Khalidov | Jonas Gehring | Zeming Lin | Jonas Gehring | Gabriel Synnaeve | Vasil Khalidov
[1] Jun Wang,et al. Multiagent Bidirectionally-Coordinated Nets for Learning to Play StarCraft Combat Games , 2017, ArXiv.
[2] Pierre Bessière,et al. A Bayesian Model for Plan Recognition in RTS Games Applied to StarCraft , 2011, AIIDE.
[3] Michael Mateas,et al. A data mining approach to strategy prediction , 2009, 2009 IEEE Symposium on Computational Intelligence and Games.
[4] Arnav Jhala,et al. A Particle Model for State Estimation in Real-Time Strategy Games , 2011, AIIDE.
[5] Chuen-Tsai Sun,et al. Building a player strategy model by analyzing replays of real-time strategy games , 2008, 2008 IEEE International Joint Conference on Neural Networks (IEEE World Congress on Computational Intelligence).
[6] Thomas G. Dietterich,et al. Inferring Strategies from Limited Reconnaissance in Real-time Strategy Games , 2012, UAI.
[7] Nicolas Usunier,et al. Episodic Exploration for Deep Deterministic Policies: An Application to StarCraft Micromanagement Tasks , 2016, ArXiv.
[8] Santiago Ontañón,et al. Automatic Learning of Combat Models for RTS Games , 2015, AIIDE.
[9] Santiago Ontañón,et al. A Survey of Real-Time Strategy Game AI Research and Competition in StarCraft , 2013, IEEE Transactions on Computational Intelligence and AI in Games.
[10] Sung-Bae Cho,et al. Replay-based strategy prediction and build order adaptation for StarCraft AI bots , 2013, 2013 IEEE Conference on Computational Inteligence in Games (CIG).
[11] Gabriel Synnaeve,et al. Programmation et apprentissage bayésien pour les jeux vidéo multi-joueurs, application à l'intelligence artificielle de jeux de stratégies temps-réel. (Bayesian Programming and Learning for Multi-Player Video Games, Application to RTS AI) , 2012 .
[12] Gabriel Synnaeve,et al. A Bayesian model for opening prediction in RTS games with application to StarCraft , 2011, 2011 IEEE Conference on Computational Intelligence and Games (CIG'11).
[13] Peng Peng,et al. Multiagent Bidirectionally-Coordinated Nets: Emergence of Human-level Coordination in Learning to Play StarCraft Combat Games , 2017, 1703.10069.
[14] Brett Browning,et al. A survey of robot learning from demonstration , 2009, Robotics Auton. Syst..
[15] Yann LeCun,et al. Deep multi-scale video prediction beyond mean square error , 2015, ICLR.
[16] Gabriel Synnaeve,et al. A Dataset for StarCraft AI and an Example of Armies Clustering , 2012 .
[17] Jee-Hyong Lee,et al. Cooperative Learning by Replay Files in Real-Time Strategy Game , 2010, CDVE.
[18] Glen Robertson,et al. An Improved Dataset and Extraction Process for Starcraft AI , 2014, FLAIRS Conference.
[19] Pieter Abbeel,et al. Apprenticeship learning via inverse reinforcement learning , 2004, ICML.
[20] Florian Richoux,et al. TorchCraft: a Library for Machine Learning Research on Real-Time Strategy Games , 2016, ArXiv.
[21] Demis Hassabis,et al. Mastering the game of Go with deep neural networks and tree search , 2016, Nature.
[22] Dorin Comaniciu,et al. Mean Shift: A Robust Approach Toward Feature Space Analysis , 2002, IEEE Trans. Pattern Anal. Mach. Intell..
[23] Shimon Whiteson,et al. Stabilising Experience Replay for Deep Multi-Agent Reinforcement Learning , 2017, ICML.