Learning multirobot joint action plans from simultaneous task execution demonstrations

The central problem of designing intelligent robot systems which learn by demonstrations of desired behaviour has been largely studied within the field of robotics. Numerous architectures for action recognition and prediction of intent of a single teacher have been proposed. However, little work has been done addressing how a group of robots can learn by simultaneous demonstrations of multiple teachers. This paper contributes a novel approach for learning multirobot joint action plans from unlabelled data. The robots firstly learn the demonstrated sequence of individual actions using the HAMMER architecture. Subsequently, the group behaviour is segmented over time and space by applying a spatio-temporal clustering algorithm. The experimental results, in which humans teleoperated real robots during a search and rescue task deployment, successfully demonstrated the efficacy of combining action recognition at individual level with group behaviour segmentation, spotting the exact moment when robots must form coalitions to achieve the goal, thus yielding reasonable generation of multirobot joint action plans.

[1]  Yiannis Demiris,et al.  Predicting the Movements of Robot Teams Using Generative Models , 2008, DARS.

[2]  K. Dautenhahn,et al.  The correspondence problem , 2002 .

[3]  Lynne E. Parker,et al.  Distributed Intelligence: Overview of the Field and Its Application in Multi-Robot Systems , 2008, AAAI Fall Symposium: Regarding the Intelligence in Distributed Intelligent Systems.

[4]  Yoram Koren,et al.  The vector field histogram-fast obstacle avoidance for mobile robots , 1991, IEEE Trans. Robotics Autom..

[5]  Michael Lewis,et al.  Human control for cooperating robot teams , 2007, 2007 2nd ACM/IEEE International Conference on Human-Robot Interaction (HRI).

[6]  Bruce A. MacDonald,et al.  Player 2.0: Toward a Practical Robot Programming Framework , 2008 .

[7]  Manuela M. Veloso,et al.  Teaching multi-robot coordination using demonstration of communication and state sharing , 2008, AAMAS.

[8]  Dieter Schmalstieg,et al.  ARToolKitPlus for Pose Trackin on Mobile Devices , 2007 .

[9]  Manuela M. Veloso,et al.  Confidence-based policy learning from demonstration using Gaussian mixture models , 2007, AAMAS '07.

[10]  Yiannis Demiris,et al.  Balancing Spectral Clustering for Segmenting Spatio-temporal Observations of Multi-agent Systems , 2008, 2008 Eighth IEEE International Conference on Data Mining.

[11]  Yiannis Demiris,et al.  Hierarchical attentive multiple models for execution and recognition of actions , 2006, Robotics Auton. Syst..

[12]  Monica N. Nicolescu,et al.  Natural methods for robot task learning: instructive demonstrations, generalization and practice , 2003, AAMAS '03.

[13]  Brett Browning,et al.  A survey of robot learning from demonstration , 2009, Robotics Auton. Syst..

[14]  Roland Siegwart,et al.  A comparison of line extraction algorithms using 2D range data for indoor mobile robotics , 2007, Auton. Robots.

[15]  Terrence Fong,et al.  Advanced Interfaces for Vehicle Teleoperation: Collaborative Control, Sensor Fusion Displays, and Remote Driving Tools , 2001, Auton. Robots.

[16]  Monica N. Nicolescu,et al.  Understanding human intentions via Hidden Markov Models in autonomous mobile robots , 2008, 2008 3rd ACM/IEEE International Conference on Human-Robot Interaction (HRI).