Engaging with Robotic Swarms

In recent years, researchers have explored human body posture and motion to control robots in more natural ways. These interfaces require the ability to track the body movements of the user in three dimensions. Deploying motion capture systems for tracking tends to be costly and intrusive and requires a clear line of sight, making them ill adapted for applications that need fast deployment. In this article, we use consumer-grade armbands, capturing orientation information and muscle activity, to interact with a robotic system through a state machine controlled by a body motion classifier. To compensate for the low quality of the information of these sensors, and to allow a wider range of dynamic control, our approach relies on machine learning. We train our classifier directly on the user to recognize (within minutes) which physiological state his or her body motion expresses. We demonstrate that on top of guaranteeing faster field deployment, our algorithm performs better than all comparable algorithms, and we detail its configuration and the most significant features extracted. As the use of large groups of robots is growing, we postulate that their interaction with humans can be eased by our approach. We identified the key factors to stimulate engagement using our system on 27 participants, each creating his or her own set of expressive motions to control a swarm of desk robots. The resulting unique dataset is available online together with the classifier and the robot control scripts.

[1]  Jian Sun,et al.  Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).

[2]  Gentiane Venture,et al.  Using Dynamics to Recognize Human Motion , 2014, Dance Notations and Robot Motion.

[3]  David St-Onge,et al.  More Than the Sum of its Parts: Assessing the Coherence and Expressivity of a Robotic Swarm , 2018, 2018 27th IEEE International Symposium on Robot and Human Interactive Communication (RO-MAN).

[4]  Lynne Anne Blom,et al.  The intimate act of choreography , 1982 .

[5]  Valérie Renaudin,et al.  Step Length Estimation Using Handheld Inertial Sensors , 2012, Sensors.

[6]  Gaël Varoquaux,et al.  Scikit-learn: Machine Learning in Python , 2011, J. Mach. Learn. Res..

[7]  F. Wilcoxon Individual Comparisons by Ranking Methods , 1945 .

[8]  Michael A. Goodrich,et al.  Haptic Shape-Based Management of Robot Teams in Cordon and Patrol , 2017, 2017 12th ACM/IEEE International Conference on Human-Robot Interaction (HRI.

[9]  Martin Molina,et al.  Natural user interfaces for human-drone multi-modal interaction , 2016, 2016 International Conference on Unmanned Aircraft Systems (ICUAS).

[10]  Salvatore Sessa,et al.  Natural human–robot musical interaction: understanding the music conductor gestures by using the WB-4 inertial measurement system , 2014, Adv. Robotics.

[11]  Chris Salter,et al.  Entangled: Technology and the Transformation of Performance , 2010 .

[12]  Guigang Zhang,et al.  Deep Learning , 2016, Int. J. Semantic Comput..

[13]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[14]  Nitish Srivastava,et al.  Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..

[15]  Pornchai Phukpattaranont,et al.  A Novel Feature Extraction for Robust EMG Pattern Recognition , 2009, ArXiv.

[16]  Frank Weichert,et al.  Analysis of the Accuracy and Robustness of the Leap Motion Controller , 2013, Sensors.

[17]  Panagiotis K. Artemiadis,et al.  A Switching Regime Model for the EMG-Based Control of a Robot Arm , 2011, IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics).

[18]  Sonia Martínez,et al.  Coverage control for mobile sensing networks , 2002, IEEE Transactions on Robotics and Automation.

[19]  Alexander Refsum Jensenius,et al.  Mumyo - evaluating and exploring the myo armband for musical interaction , 2015, NIME.

[20]  Yanjun Qi Random Forest for Bioinformatics , 2012 .

[21]  Marco Dorigo,et al.  Investigating the effect of increasing robot group sizes on the human psychophysiological state in the context of human–swarm interaction , 2016, Swarm Intelligence.

[22]  Erik Scheme,et al.  Electromyogram pattern recognition for control of powered upper-limb prostheses: state of the art and challenges for clinical use. , 2011, Journal of rehabilitation research and development.

[23]  Hiroshi Sato,et al.  Multiple targets enclosure by robotic swarm , 2014, Robotics Auton. Syst..

[24]  Clément Gosselin,et al.  A convolutional neural network for robotic arm guidance using sEMG based frequency-features , 2016, 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[25]  Mithileysh Sathiyanarayanan,et al.  MYO Armband for physiotherapy healthcare: A case study using gesture recognition application , 2016, 2016 8th International Conference on Communication Systems and Networks (COMSNETS).

[26]  Federico Manuri,et al.  A Kinect-based natural interface for quadrotor control , 2011, Entertain. Comput..

[27]  Carlo Pinciroli,et al.  Swarm-Oriented Programming of Distributed Robot Networks , 2016, Computer.

[28]  Gregory Dudek,et al.  A Simple Tactile Probe for Surface Identification by Mobile Robots , 2011, IEEE Transactions on Robotics.

[29]  Reid G. Simmons,et al.  Keep on dancing: Effects of expressive motion mimicry , 2017, 2017 26th IEEE International Symposium on Robot and Human Interactive Communication (RO-MAN).

[30]  Pierre Dragicevic,et al.  Zooids: Building Blocks for Swarm User Interfaces , 2016, UIST.

[31]  Zhaozheng Yin,et al.  Human Activity Recognition Using Wearable Sensors by Deep Convolutional Neural Networks , 2015, ACM Multimedia.

[32]  Soon Mook Jung,et al.  The development of interface device for human robot interaction , 2007, 2007 International Conference on Control, Automation and Systems.

[33]  Jürgen Schmidhuber,et al.  Long Short-Term Memory , 1997, Neural Computation.

[34]  Jake K. Aggarwal,et al.  Human Motion Analysis: A Review , 1999, Comput. Vis. Image Underst..

[35]  Luca Antiga,et al.  Automatic differentiation in PyTorch , 2017 .

[36]  Angela Loureiro de Souza Laban Movement Analysis - Scaffolding Human Movement to Multiply Possibilities and Choices , 2014, Dance Notations and Robot Motion.

[37]  Bill Vorn I Want to Believe—Empathy and Catharsis in Robotic Art , 2016 .

[38]  Gideon Steinberg Natural User Interfaces , 2012 .

[39]  Martin A. Giese,et al.  Bayesian Approaches for Learning of Primitive-Based Compact Representations of Complex Human Activities , 2014, Dance Notations and Robot Motion.

[40]  Atsuo Takanishi,et al.  Musical-based interaction system for the Waseda Flutist Robot , 2010, Auton. Robots.

[41]  Andrea Gasparri,et al.  Decentralized progressive shape formation with robot swarms , 2018, Autonomous Robots.

[42]  Sergey Ioffe,et al.  Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift , 2015, ICML.

[43]  Elisabetta Zibetti,et al.  Interacting with Non-anthropomorphic Robotic Artworks and Interpreting Their Behaviour , 2017, Int. J. Soc. Robotics.

[44]  A. Phinyomark,et al.  Evaluation of EMG feature extraction for hand movement recognition based on Euclidean distance and standard deviation , 2010, ECTI-CON2010: The 2010 ECTI International Confernce on Electrical Engineering/Electronics, Computer, Telecommunications and Information Technology.

[45]  Toshio Tsuji,et al.  A Hybrid Motion Classification Approach for EMG-Based Human–Robot Interfaces Using Bayesian and Neural Networks , 2009, IEEE Transactions on Robotics.

[46]  Gregory Dudek,et al.  A natural gesture interface for operating robotic systems , 2008, 2008 IEEE International Conference on Robotics and Automation.

[47]  Michael A. Goodrich,et al.  Two invariants of human-swarm interaction , 2015, Journal of Human-Robot Interaction.

[48]  Katia P. Sycara,et al.  Neglect benevolence in human control of swarms in the presence of latency , 2012, 2012 IEEE International Conference on Systems, Man, and Cybernetics (SMC).

[49]  Timothy H. Lucas,et al.  The Virtual Trackpad: An Electromyography-Based, Wireless, Real-Time, Low-Power, Embedded Hand-Gesture-Recognition System Using an Event-Driven Artificial Neural Network , 2017, IEEE Transactions on Circuits and Systems II: Express Briefs.

[50]  P. Ekman Are there basic emotions? , 1992, Psychological review.

[51]  Hyeran Byun,et al.  Human tracking and silhouette extraction for human–robot interaction systems , 2008, Pattern Analysis and Applications.

[52]  Lawrence H. Kim,et al.  Human Perception of Swarm Robot Motion , 2017, CHI Extended Abstracts.

[53]  Damith C. Herath,et al.  Robots and Art: Exploring an Unlikely Symbiosis , 2016 .

[54]  Sergio A. Velastin,et al.  A Kinect-based 3D hand-gesture interface for 3D databases , 2014, Journal on Multimodal User Interfaces.

[55]  Tony J. Dodd,et al.  Evolving Aggregation Behaviors in Multi-Robot Systems with Binary Sensors , 2012, DARS.

[56]  Michael A. Goodrich,et al.  Scalable Human Interaction with Robotic Swarms , 2013 .

[57]  Zoubin Ghahramani,et al.  Dropout as a Bayesian Approximation: Representing Model Uncertainty in Deep Learning , 2015, ICML.

[58]  Klaus Bengler,et al.  A questionnaire for the evaluation of physical assistive devices (QUEAD): Testing usability and acceptance in physical human-robot interaction , 2017, 2017 IEEE International Conference on Systems, Man, and Cybernetics (SMC).

[59]  Yunming Ye,et al.  Classifying Very High-Dimensional Data with Random Forests Built from Small Subspaces , 2012, Int. J. Data Warehous. Min..

[60]  Rocco J. Perla,et al.  Ten Common Misunderstandings, Misconceptions, Persistent Myths and Urban Legends about Likert Scales and Likert Response Formats and their Antidotes , 2007 .

[61]  Lawrence H. Kim,et al.  UbiSwarm: Ubiquitous Robotic Interfaces and Investigation of Abstract Motion as a Display , 2017, Proc. ACM Interact. Mob. Wearable Ubiquitous Technol..

[62]  Lyuba Alboul,et al.  A Robot Swarm Assisting a Human Fire-Fighter , 2009, Adv. Robotics.

[63]  Carlo Ghezzi,et al.  Team-level programming of drone sensor networks , 2014, SenSys.

[64]  S. Rigatti Random Forest. , 2017, Journal of insurance medicine.

[65]  Clément Gosselin,et al.  Deep Learning for Electromyographic Hand Gesture Signal Classification Using Transfer Learning , 2018, IEEE Transactions on Neural Systems and Rehabilitation Engineering.

[66]  Katia P. Sycara,et al.  Human Interaction With Robot Swarms: A Survey , 2016, IEEE Transactions on Human-Machine Systems.

[67]  Ho-Sub Yoon,et al.  Vision-based arm gesture recognition for a long-range human–robot interaction , 2010, The Journal of Supercomputing.

[68]  Panagiotis K. Artemiadis,et al.  Task discrimination from myoelectric activity: A learning scheme for EMG-based interfaces , 2013, 2013 IEEE 13th International Conference on Rehabilitation Robotics (ICORR).