Defining and using ideal teammate and opponent agent models: a case study in robotic soccer

A common challenge for agents in multiagent systems is trying to predict what other agents are going to do in the future. Such knowledge can help an agent determine which of its current action options are most likely to help it achieve its goals. Ideally, an agent could learn a model of other agents' behavior patterns via direct observation of their past actions. However, that is only possible when agents have many repeated interactions with one another. We explore the use of agent models in an application where extensive interactions with a particular agent are not possible, namely robotic soccer. In robotic soccer tournaments, such as RoboCup (Kitano et al., 1997), a team of agents plays against another team for a single, short (typically 10-minute) period. The opponents' behaviors are usually not observable prior to this game and there are not enough interactions during the game to build a useful model. We introduce "ideal-model-based behavior outcome prediction" (IMBBOP). This technique predicts an agent's future actions in relation to the optimal behavior in its given situation, This optimal behavior is agent-independent and can therefore be computed based solely on a model of the world dynamics. IMBBOP does not assume that the other agent will act according to the theoretical optimum, but rather characterizes its expected behavior in terms of deviation from this optimum.

[1]  Edmund H. Durfee,et al.  A decision-theoretic approach to coordinating multiagent interactions , 1991, IJCAI 1991.

[2]  Edmund H. Durfee,et al.  Blissful Ignorance: Knowing Just Enough to Coordinate Well , 1995, ICMAS.

[3]  Edmund H. Durfee,et al.  Recursive Agent Modeling Using Limited Rationality , 1995, ICMAS.

[4]  Milind Tambe Recursive Agent and Agent-Group Tracking in a Real-Time Dynamic Environment , 1995, ICMAS.

[5]  Edmund H. Durfee,et al.  A Rigorous, Operational Formalization of Recursive Modeling , 1995, ICMAS.

[6]  Edmund H. Durfee,et al.  Deciding When to Commit to Action During Observation-Based Coordination , 1995, ICMAS.

[7]  Xuemei Wang Planning While Learning Operators , 1996, AIPS.

[8]  Hiroaki Kitano,et al.  The RoboCup Synthetic Agent Challenge 97 , 1997, IJCAI.

[9]  Peter Stone,et al.  Layered Learning in Multiagent Systems , 1997, AAAI/IAAI.

[10]  Ian Frank,et al.  Soccer Server: A Tool for Research on Multiagent Systems , 1998, Appl. Artif. Intell..

[11]  Manuela M. Veloso,et al.  The CMUnited-98 Champion Simulator Team , 1998, RoboCup.

[12]  浅田 稔,et al.  RoboCup-98 : Robot Soccer World Cup II , 1999 .

[13]  Manuela M. Veloso,et al.  The CMUnited-99 Champion Simulator Team , 2000, AI Mag..

[14]  Peter Stone,et al.  Layered learning in multiagent systems - a winning approach to robotic soccer , 2000, Intelligent robotics and autonomous agents.

[15]  D. O A L O N S O,et al.  Learning in multi-agent systems , 2002 .