Evolving Game Skill-Depth using General Video Game AI agents

Most games have, or can be generalised to have, a number of parameters that may be varied in order to provide instances of games that lead to very different player experiences. The space of possible parameter settings can be seen as a search space, and we can therefore use a Random Mutation Hill Climbing algorithm or other search methods to find the parameter settings that induce the best games. One of the hardest parts of this approach is defining a suitable fitness function. In this paper we explore the possibility of using one of a growing set of General Video Game AI agents to perform automatic play-testing. This enables a very general approach to game evaluation based on estimating the skill-depth of a game. Agent-based play-testing is computationally expensive, so we compare two simple but efficient optimisation algorithms: the Random Mutation Hill-Climber and the Multi-Armed Bandit Random Mutation Hill-Climber. For the test game we use a space-battle game in order to provide a suitable balance between simulation speed and potential skill-depth. Results show that both algorithms are able to rapidly evolve game versions with significant skill-depth, but that choosing a suitable resampling number is essential in order to combat the effects of noise.

[1]  Simon M. Lucas,et al.  Optimal resampling for the noisy OneMax problem , 2016, ArXiv.

[2]  J. Togelius,et al.  Discovering Unique Game Variants , 2015 .

[3]  Michael Mateas,et al.  Towards Automated Game Design , 2007, AI*IA.

[4]  Julian Togelius,et al.  Search-Based Procedural Content Generation: A Taxonomy and Survey , 2011, IEEE Transactions on Computational Intelligence and AI in Games.

[5]  Simon Colton,et al.  Multi-faceted evolution of simple arcade games , 2011, 2011 IEEE Conference on Computational Intelligence and Games (CIG'11).

[6]  Jialin Liu Portfolio methods in uncertain contexts , 2015 .

[7]  Julian Togelius,et al.  An experiment in automatic game design , 2008, 2008 IEEE Symposium On Computational Intelligence and Games.

[8]  Julian Togelius,et al.  Depth in Strategic Games , 2017, AAAI Workshops.

[9]  Philippe Pasquier,et al.  Towards a Generic Framework for Automated Video Game Level Creation , 2010, EvoApplications.

[10]  Julian Togelius,et al.  Towards automatic personalised content creation for racing games , 2007, 2007 IEEE Symposium on Computational Intelligence and Games.

[11]  Simon M. Lucas,et al.  The N-Tuple bandit evolutionary algorithm for automatic game improvement , 2017, 2017 IEEE Congress on Evolutionary Computation (CEC).

[12]  Simon M. Lucas,et al.  Learning DFA: evolution versus evidence driven state merging , 2003, The 2003 Congress on Evolutionary Computation, 2003. CEC '03..

[13]  Simon M. Lucas,et al.  Rolling Horizon Coevolutionary planning for two-player video games , 2016, 2016 8th Computer Science and Electronic Engineering (CEEC).

[14]  Kenneth O. Stanley,et al.  Automatic Content Generation in the Galactic Arms Race Video Game , 2009, IEEE Transactions on Computational Intelligence and AI in Games.

[15]  Daniel A. Ashlock,et al.  Automatic generation of game elements via evolution , 2010, Proceedings of the 2010 IEEE Conference on Computational Intelligence and Games.

[16]  Simon Colton,et al.  Aesthetic Considerations for Automated Platformer Design , 2012, AIIDE.

[17]  Julian Togelius,et al.  General Video Game Evaluation Using Relative Algorithm Performance Profiles , 2015, EvoApplications.

[18]  Simon M. Lucas,et al.  Bandit-based Random Mutation Hill-Climbing , 2016, 2017 IEEE Congress on Evolutionary Computation (CEC).

[19]  Simon M. Lucas,et al.  Learning deterministic finite automata with a smart state labeling evolutionary algorithm , 2005, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[20]  Simon Colton,et al.  Automated tweaking of levels for casual creation of mobile games , 2016 .

[21]  Frédéric Maire,et al.  Evolutionary Game Design , 2011, IEEE Transactions on Computational Intelligence and AI in Games.