M-ROSE: A Multi Robot Simulation Environment for Learning Cooperative Behavior

The development of high-performance autonomous multi robot control systems requires intensive experimentation in controllable, repeatable, and realistic robot settings. The need for experimentation is even higher in applications where the robots should automatically learn substantial parts of their controllers. We propose to solve such learning tasks as a three step process. First, we learn a simulator of the robots’ dynamics. Second, we perform the learning tasks using the learned simulator. Third, we port the learned controller to the real robot and cross validate the performance gains obtained by the learned controllers. In this paper, we describe M-ROSE, our learning simulator, and provide empirical evidence that it is a powerful tool for learning of sophisticated control modules for real robots.

[1]  Michael Beetz,et al.  Learning to Execute Navigation Plans , 2001, KI/ÖGAI.

[2]  Rodney A. Brooks,et al.  Artificial Life and Real Robots , 1992 .

[3]  Rodney A. Brooks,et al.  Real Robots, Real Learning Problems , 1993 .

[4]  Martin A. Riedmiller,et al.  Karlsruhe Brainstormers - A Reinforcement Learning Approach to Robotic Soccer , 2000, RoboCup.

[5]  Ian Frank,et al.  Soccer Server: A Tool for Research on Multiagent Systems , 1998, Appl. Artif. Intell..

[6]  Inman Harvey,et al.  Noise and the Reality Gap: The Use of Simulation in Evolutionary Robotics , 1995, ECAL.

[7]  Michael Beetz,et al.  The AGILO autonomous robot soccer team: computational principles, experiences, and perspectives , 2002, AAMAS '02.

[8]  Roger J. Hubbold,et al.  Mobile Robot Simulation by Means of Acquired Neural Network Models , 1998, ESM.

[9]  A. Sydow,et al.  Parallelity in high-level simulation architectures , 1998 .

[10]  Michael Beetz,et al.  Reliable Multi-robot Coordination Using Minimal Communication and Neural Prediction , 2001, Advances in Plan-Based Control of Robotic Agents.

[11]  Wolfram Burgard,et al.  Map learning and high-speed navigation in RHINO , 1998 .

[12]  Michael Beetz,et al.  Multi-robot path planning for dynamic environments: a case study , 2001, Proceedings 2001 IEEE/RSJ International Conference on Intelligent Robots and Systems. Expanding the Societal Role of Robotics in the the Next Millennium (Cat. No.01CH37180).

[13]  Martin A. Riedmiller,et al.  A direct adaptive method for faster backpropagation learning: the RPROP algorithm , 1993, IEEE International Conference on Neural Networks.

[14]  R. Palmer,et al.  Introduction to the theory of neural computation , 1994, The advanced book program.

[15]  Sridhar Mahadevan,et al.  Robot Learning , 1993 .