Distributed Algorithms for Multi-Robot Observation of Multiple Moving Targets

An important issue that arises in the automation of many security, surveillance, and reconnaissance tasks is that of observing the movements of targets navigating in a bounded area of interest. A key research issue in these problems is that of sensor placement—determining where sensors should be located to maintain the targets in view. In complex applications involving limited-range sensors, the use of multiple sensors dynamically moving over time is required. In this paper, we investigate the use of a cooperative team of autonomous sensor-based robots for the observation of multiple moving targets. In other research, analytical techniques have been developed for solving this problem in complex geometrical environments. However, these previous approaches are very computationally expensive—at least exponential in the number of robots—and cannot be implemented on robots operating in real-time. Thus, this paper reports on our studies of a simpler problem involving uncluttered environments—those with either no obstacles or with randomly distributed simple convex obstacles. We focus primarily on developing the on-line distributed control strategies that allow the robot team to attempt to minimize the total time in which targets escape observation by some robot team member in the area of interest. This paper first formalizes the problem (which we term CMOMMT for Cooperative Multi-Robot Observation of Multiple Moving Targets) and discusses related work. We then present a distributed heuristic approach (which we call A-CMOMMT) for solving the CMOMMT problem that uses weighted local force vector control. We analyze the effectiveness of the resulting weighted force vector approach by comparing it to three other approaches. We present the results of our experiments in both simulation and on physical robots that demonstrate the superiority of the A-CMOMMT approach for situations in which the ratio of targets to robots is greater than 1/2. Finally, we conclude by proposing that the CMOMMT problem makes an excellent domain for studying multi-robot learning in inherently cooperative tasks. This approach is the first of its kind for solving the on-line cooperative observation problem and implementing it on a physical robot team.

[1]  Leonidas J. Guibas,et al.  Finding an unpredictable target in a workspace with obstacles , 1997, Proceedings of International Conference on Robotics and Automation.

[2]  Cathleen Stasz,et al.  Network Structures for Distributed Situation Assessment , 1980, IEEE Transactions on Systems, Man, and Cybernetics.

[3]  Sandip Sen,et al.  Evolution and learning in multiagent systems , 1998, Int. J. Hum. Comput. Stud..

[4]  Robert M. Gray,et al.  An Algorithm for Vector Quantizer Design , 1980, IEEE Trans. Commun..

[5]  Yaser Al-Onaizan,et al.  On being a teammate: experiences acquired in the design of RoboCup teams , 1999, AGENTS '99.

[6]  Yaakov Bar-Shalom,et al.  Tracking methods in a multitarget environment , 1978 .

[7]  M. Benda,et al.  On Optimal Cooperation of Knowledge Sources , 1985 .

[8]  Gerhard Weiß,et al.  Adaptation and Learning in Multi-Agent Systems: Some Remarks and a Bibliography , 1995, Adaption and Learning in Multi-Agent Systems.

[9]  L. E. ParkerCenter Learning in Large Cooperative Multi-Robot Domains , 2001 .

[10]  Lynne E. Parker,et al.  Multi-Robot Learning in a Cooperative Observation Task , 2000, DARS.

[11]  Lynne E. Parker Distributed control of multi-robot teams: Cooperative baton passing task , 1998 .

[12]  Yaakov Bar-Shalom,et al.  Multitarget-multisensor tracking: Advanced applications , 1989 .

[13]  Lynne E. Parker,et al.  ALLIANCE: an architecture for fault tolerant, cooperative control of heterogeneous mobile robots , 1994 .

[14]  Lynne E. Parker An Experiment in Mobile Robotic Cooperation , 1994 .

[15]  Edmund H. Durfee,et al.  Coherent Cooperation Among Communicating Problem Solvers , 1987, IEEE Transactions on Computers.

[16]  J. O'Rourke Art gallery theorems and algorithms , 1987 .

[17]  Frederick Hayes-Roth,et al.  Distributed Intelligence for Air Fleet Control , 1981 .

[18]  Fernando Fernández,et al.  VQQL. Applying Vector Quantization to Reinforcement Learning , 1999, RoboCup.

[19]  Yaakov Bar-Shalom,et al.  Multitarget-Multisensor Tracking , 1995 .

[20]  Geoffrey C. Fox,et al.  Parallel Computing Works , 1994 .

[21]  S. Griffis EDITOR , 1997, Journal of Navigation.

[22]  S. P. Lloyd,et al.  Least squares quantization in PCM , 1982, IEEE Trans. Inf. Theory.

[23]  Sandip Sen,et al.  Adaption and Learning in Multi-Agent Systems , 1995, Lecture Notes in Computer Science.

[24]  C. R. Henson Conclusion , 1969 .

[25]  Maja J. Mataric,et al.  Interaction and intelligent behavior , 1994 .

[26]  Lynne E. Parker,et al.  Cooperative Robotics for Multi-Target Observation , 1999, Intell. Autom. Soft Comput..

[27]  Masafumi Yamashita,et al.  The Searchlight Scheduling Problem , 1990, SIAM J. Comput..

[28]  P. Bernhard,et al.  Saddle point conditions for a class of stochastic dynamical games with imperfect information , 1988 .

[29]  Sandip Sen,et al.  Evolving Beharioral Strategies in Predators and Prey , 1995, Adaption and Learning in Multi-Agent Systems.

[30]  Lynne E. Parker,et al.  On the design of behavior-based multi-robot teams , 1995, Adv. Robotics.

[31]  Masafumi Yamashita,et al.  Searching for a Mobile Intruder in a Polygonal Region , 1992, SIAM J. Comput..

[32]  Héctor H. González-Baños,et al.  Motion strategies for maintaining visibility of a moving target , 1997, Proceedings of International Conference on Robotics and Automation.

[33]  Hiroaki Kitano,et al.  RoboCup: A Challenge Problem for AI and Robotics , 1997, RoboCup.

[34]  Lynne E. Parker Multi-Robot Team Design for Real-World Applications , 1996 .

[35]  Amy J. Briggs Efficient Geometric Algorithms for Robot Sensing and Control , 1995 .

[36]  Hobart R. Everett,et al.  Coordinated control of multiple security robots , 1994, Other Conferences.

[37]  Sridhar Mahadevan,et al.  Automatic Programming of Behavior-Based Robots Using Reinforcement Learning , 1991, Artif. Intell..

[38]  Lynne E. Parker,et al.  ALLIANCE: an architecture for fault tolerant multirobot cooperation , 1998, IEEE Trans. Robotics Autom..

[39]  Manuela M. Veloso,et al.  Layered Approach to Learning Client Behaviors in the Robocup Soccer Server , 1998, Appl. Artif. Intell..

[40]  David W. Aha,et al.  Lazy Learning , 1997, Springer Netherlands.

[41]  P. Bernhard Computation of equilibrium points of delayed partial information pursuit evasion games , 1987, 26th IEEE Conference on Decision and Control.