Real-time recognition of team behaviors by multisensory graph-embedded robot learning

Awareness of team behaviors (e.g., individual activities and team intents) plays a critical role in human–robot teaming. Autonomous robots need to be aware of the overall intent of the team they are collaborating with in order to effectively aid their human peers or augment the team’s capabilities. Team intents encode the goal of the team, which cannot be simply identified from a collection of individual activities. Instead, teammate relationships must also be encoded for team intent recognition. In this article, we introduce a novel representation learning approach to recognizing team intent awareness in real-time, based upon both individual human activities and the relationship between human peers in the team. Our approach formulates the task of robot learning for team intent recognition as a joint regularized optimization problem, which encodes individual activities as latent variables and represents teammate relationships through graph embedding. In addition, we design a new algorithm to efficiently solve the formulated regularized optimization problem, which possesses a theoretical guarantee to converge to the optimal solution. To evaluate our approach’s performance on team intent recognition, we test our approach on a public benchmark group activity dataset and a multisensory underground search and rescue team behavior dataset newly collected from robots in an underground environment, as well as perform a proof-of-concept case study on a physical robot. The experimental results have demonstrated both the superior accuracy of our proposed approach and its suitability for real-time applications on mobile robots.

[1]  Peter Stone,et al.  Autonomous agents modelling other agents: A comprehensive survey and open problems , 2017, Artif. Intell..

[2]  Piotr J. Gmytrasiewicz,et al.  Interactive POMDPs with finite-state models of other agents , 2017, Autonomous Agents and Multi-Agent Systems.

[3]  Yang Wang,et al.  Discriminative Latent Models for Recognizing Contextual Group Activities , 2012, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[4]  Demis Hassabis,et al.  Mastering the game of Go with deep neural networks and tree search , 2016, Nature.

[5]  Paul Robinette,et al.  Preliminary Interactions of Human-Robot Trust, Cognitive Load, and Robot Intelligence Levels in a Competitive Game , 2018, HRI.

[6]  Christopher Crick,et al.  Learning to Assess the Cognitive Capacity of Human Partners , 2017, HRI.

[7]  A. W. Evans,et al.  Communicating intent to develop shared situation awareness and engender trust in human-agent teams , 2017, Cognitive Systems Research.

[8]  Marc Toussaint,et al.  Preferred Interaction Styles for Human-Robot Collaboration Vary Over Tasks With Different Action Types , 2018, Front. Neurorobot..

[9]  Yu Zhang,et al.  A human factors analysis of proactive support in human-robot teaming , 2015, 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[10]  Joseph J. LaViola,et al.  Measuring and reducing observational latency when recognizing actions , 2011, 2011 IEEE International Conference on Computer Vision Workshops (ICCV Workshops).

[11]  Julie A. Shah,et al.  Computational design of mixed-initiative human–robot teaming that considers human factors: situational awareness, workload, and workflow preferences , 2017, Int. J. Robotics Res..

[12]  Silvio Savarese,et al.  Understanding Collective Activitiesof People from Videos , 2014, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[13]  Junsong Yuan,et al.  Learning Actionlet Ensemble for 3D Human Action Recognition , 2014, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[14]  Fei Han,et al.  Space-Time Representation of People Based on 3D Skeletal Data: A Review , 2016, Comput. Vis. Image Underst..

[15]  Jake K. Aggarwal,et al.  Spatio-temporal relationship match: Video structure comparison for recognition of complex human activities , 2009, 2009 IEEE 12th International Conference on Computer Vision.

[16]  Araceli Sanchis,et al.  OMBO: An opponent modeling approach , 2009, AI Commun..

[17]  Stefanos Nikolaidis,et al.  Improved human–robot team performance through cross-training, an approach inspired by human team training practices , 2015, Int. J. Robotics Res..

[18]  Roland Siegwart,et al.  Experience in system design for human-robot teaming in urban search & rescue , 2014 .

[19]  Xue Yang,et al.  SRAL: Shared Representative Appearance Learning for Long-Term Visual Place Recognition , 2017, IEEE Robotics and Automation Letters.

[20]  Julie A. Adams,et al.  A Diagnostic Human Workload Assessment Algorithm for Human-Robot Teams , 2018, HRI.

[21]  Kevin Lee,et al.  Come See This! Augmented Reality to Enable Human-Robot Cooperative Search , 2018, 2018 IEEE International Symposium on Safety, Security, and Rescue Robotics (SSRR).

[22]  Sung-Bae Cho,et al.  Human activity recognition with smartphone sensors using deep learning neural networks , 2016, Expert Syst. Appl..

[23]  Michal Król,et al.  Wireless Sensor Networks and Multi-UAV systems for natural disaster management , 2017, Comput. Networks.

[24]  Daniel Barber,et al.  Achieving the Vision of Effective Soldier-Robot Teaming: Recent Work in Multimodal Communication , 2015, HRI.

[25]  Jonathan L. Shapiro,et al.  Opponent Modeling by Expectation–Maximization and Sequence Prediction in Simplified Poker , 2017, IEEE Transactions on Computational Intelligence and AI in Games.

[26]  Lynne E. Parker,et al.  Skeleton-based bio-inspired human activity prediction for real-time human–robot interaction , 2017, Autonomous Robots.

[27]  Mohamed R. Amer,et al.  HiRF: Hierarchical Random Field for Collective Activity Recognition in Videos , 2014, ECCV.

[28]  Daniel Roggen,et al.  Deep Convolutional and LSTM Recurrent Neural Networks for Multimodal Wearable Activity Recognition , 2016, Sensors.