Integrating Case-Based Reasoning with Reinforcement Learning for Real-Time Strategy Game Micromanagement

This paper describes the conception of a hybrid Reinforcement Learning (RL) and Case-Based Reasoning (CBR) approach to managing combat units in strategy games. Both methods are combined into an AI agent that is evaluated by using the real-time strategy (RTS) computer game StarCraft as a test bed. The eventual aim of this approach is an AI agent that has the same actions and information at its disposal as a human player. As part of an experimental evaluation, the agent is tested in different scenarios using optimized algorithm parameters. The integration of CBR for memory management is shown to improve the speed of convergence to an optimal policy, while also enabling the agent to address a larger variety of problems when compared to simple RL. The agent manages to beat the built-in game AI and also outperforms a simple RL-only agent. An analysis of the evolution of the case-base shows how scenarios and algorithmic parameters influence agent performance and will serve as a foundation for future improvement to the hybrid CBR/RL approach.

[1]  Babak Esfandiari,et al.  Use of Fuzzy Histograms to Model the Spatial Distribution of Objects in Case-Based Reasoning , 2008, Canadian Conference on AI.

[2]  Santiago Ontañón,et al.  Kiting in RTS Games Using Influence Maps , 2012, Proceedings of the AAAI Conference on Artificial Intelligence and Interactive Digital Entertainment.

[3]  Ian D. Watson,et al.  Applying reinforcement learning to small scale combat in the real-time strategy game StarCraft:Broodwar , 2012, 2012 IEEE Conference on Computational Intelligence and Games (CIG).

[4]  Michael Buro,et al.  Portfolio greedy search and simulation for large-scale combat in starcraft , 2013, 2013 IEEE Conference on Computational Inteligence in Games (CIG).

[5]  Stefan Wess,et al.  Using k-d Trees to Improve the Retrieval Step in Case-Based Reasoning , 1993, EWCBR.

[6]  Agnar Aamodt,et al.  Case-Based Reasoning: Foundational Issues, Methodological Variations, and System Approaches , 1994, AI Commun..

[7]  Arnav Jhala,et al.  Applying Goal-Driven Autonomy to StarCraft , 2010, AIIDE.

[8]  Aleksandar Micic,et al.  Developing game AI for the real‐time strategy game StarCraft , 2011 .

[9]  Johan Hagelbäck,et al.  Multi-Agent Potential Field Based Architectures for Real-Time Strategy Game Bots , 2012 .

[10]  Michael Buro,et al.  Fast Heuristic Search for RTS Game Combat Scenarios , 2012, AIIDE.

[11]  Chris Watkins,et al.  Learning from delayed rewards , 1989 .

[12]  Arnav Jhala,et al.  Building Human-Level AI for Real-Time Strategy Games , 2011, AAAI Fall Symposium: Advances in Cognitive Systems.

[13]  Stefan Wess,et al.  Topics in Case-Based Reasoning , 1994 .

[14]  Richard S. Sutton,et al.  Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.

[15]  O. Khatib,et al.  Real-Time Obstacle Avoidance for Manipulators and Mobile Robots , 1985, Proceedings. 1985 IEEE International Conference on Robotics and Automation.

[16]  E. M. Wright,et al.  Adaptive Control Processes: A Guided Tour , 1961, The Mathematical Gazette.

[17]  Michael Buro,et al.  Build Order Optimization in StarCraft , 2011, AIIDE.

[18]  David W. Aha,et al.  Learning Continuous Action Models in a Real-Time Strategy Environment , 2008, FLAIRS.

[19]  Santiago Ontañón,et al.  A Survey of Real-Time Strategy Game AI Research and Competition in StarCraft , 2013, IEEE Transactions on Computational Intelligence and AI in Games.

[20]  Michael Buro,et al.  RTS Games and Real-Time AI Research , 2003 .