Evolving explicit opponent models in game playing

Opponent models are necessary in games where the game state is only partially known to the player, since the player must infer the state of the game based on the opponents actions. This paper presents an architecture and a process for developing neural network game players that utilize explicit opponent models in order to improve game play against unseen opponents. The model is constructed as a mixture over a set of cardinal opponents, i.e. opponents that represent maximally distinct game strategies. The model is trained to estimate the likelihood that the opponent will make the same move as each of the cardinal opponents would in a given game situation. Experiments were performed in the game of Guess It, a simple game of imperfect information that has no optimal strategy for defeating specific opponents. Opponent modeling is therefore crucial to play this game well. Both opponent modeling and game-playing neural networks were trained using NeuroEvolution of Augmenting Topologies (NEAT). The results demonstrate that game-playing provided with the model outperform networks not provided with the model when played against the same previously unseen opponents. The cardinal mixture architecture therefore constitutes a promising approach for general and dynamic opponent modeling in game-playing.

[1]  Eric O. Postma,et al.  Adaptive game AI with dynamic scripting , 2006, Machine Learning.

[2]  Manuela M. Veloso,et al.  Planning for Distributed Execution through Use of Probabilistic Opponent Models , 2002, AIPS.

[3]  Risto Miikkulainen,et al.  Evolving robocup keepaway players through task decomposition , 2003 .

[4]  Risto Miikkulainen,et al.  Continual Coevolution Through Complexification , 2002, GECCO.

[5]  Rufus Isaacs,et al.  A Card Game with Bluffing , 1955 .

[6]  M. Ponsen Automatically Generating Game Tactics via Evolutionary Learning , 2005 .

[7]  David W. Aha,et al.  Automatically Generating Game Tactics through Evolutionary Learning , 2006, AI Mag..

[8]  Jonathan Schaeffer,et al.  Opponent Modeling in Poker , 1998, AAAI/IAAI.

[9]  Risto Miikkulainen,et al.  Evolving Soccer Keepaway Players Through Task Decomposition , 2005, Machine Learning.

[10]  Teuvo Kohonen,et al.  The self-organizing map , 1990, Neurocomputing.

[11]  Kevin B. Korb,et al.  Bayesian Poker , 1999, UAI.

[12]  Michael H. Bowling,et al.  Bayes' Bluff: Opponent Modelling in Poker , 2005, UAI 2005.

[13]  Risto Miikkulainen,et al.  Evolving Keepaway Soccer Players through Task Decomposition , 2003, GECCO.

[14]  Jonathan Schaeffer,et al.  Improved Opponent Modeling in Poker , 2000 .

[15]  Risto Miikkulainen,et al.  2-D Pole Balancing with Recurrent Evolutionary Networks , 1998 .

[16]  R. Lyndon While,et al.  Learning In RoboCup Keepaway Using Evolutionary Algorithms , 2002, GECCO.