Biomechanical-Based Approach to Data Augmentation for One-Shot Gesture Recognition

Most common approaches to one-shot gesture recognition have leveraged mainly conventional machine learning solutions and image based data augmentation techniques, ignoring the mechanisms that are used by humans to perceive and execute gestures, a key contextual component in this process. The novelty of this work consists on modeling the process that leads to the creation of gestures, rather than observing the gesture alone. In this approach, the context considered involves the way in which humans produce the gestures – the kinematic and biomechanical characteristics associated with gesture production and execution. By understanding the main "modes" of variation we can replicate the single observation many times. Consequently, the main strategy proposed in this paper includes generating a data set of human-like examples based on "naturalistic" features extracted from a single gesture sample while preserving fundamentally human characteristics like visual saliency, smooth transitions and economy of motion. The availability of a large data set of realistic samples allows the use state-of-the-art classifiers for further recognition. Several classifiers were trained and their recognition accuracies were assessed and compared to previous one-shot learning approaches. An average recognition accuracy of 95% among all classifiers highlights the relevance of keeping the human "in the loop" to effectively achieve one-shot gesture recognition.

[1]  L. Deng,et al.  The MNIST Database of Handwritten Digit Images for Machine Learning Research [Best of the Web] , 2012, IEEE Signal Processing Magazine.

[2]  Alexandre Barachant,et al.  Using Data From the Microsoft Kinect 2 to Quantify Upper Limb Behavior: A Feasibility Study , 2017, IEEE Journal of Biomedical and Health Informatics.

[3]  A. Torres Ruiz,et al.  Biomechanical Characterization of Five Sporting Gestures of Taekwondo , 2015 .

[4]  Giorgio Metta,et al.  Keep it simple and sparse: real-time action recognition , 2013, J. Mach. Learn. Res..

[5]  Oriol Vinyals,et al.  Matching Networks for One Shot Learning , 2016, NIPS.

[6]  Joseph J. LaViola,et al.  Exploring the Trade-off Between Accuracy and Observational Latency in Action Recognition , 2013, International Journal of Computer Vision.

[7]  Panagiotis Artemiadis,et al.  Closed-form Inverse Kinematic Solution for Anthropomorphic Motion in Redundant Robot Arms , 2013, ICRA 2013.

[8]  Juan Pablo Wachs,et al.  What Makes a Gesture a Gesture? Neural Signatures Involved in Gesture Recognition , 2017, 2017 12th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2017).

[9]  Nicolas D. Georganas,et al.  Real-Time Hand Gesture Detection and Recognition Using Bag-of-Features and Support Vector Machine Techniques , 2011, IEEE Transactions on Instrumentation and Measurement.

[10]  Richard P. Paul,et al.  Robot manipulators : mathematics, programming, and control : the computer control of robot manipulators , 1981 .

[11]  Rajesh B. Mapari,et al.  Real time human pose recognition using leap motion sensor , 2015, 2015 IEEE International Conference on Research in Computational Intelligence and Communication Networks (ICRCICN).

[12]  Paolo Dario,et al.  A biomechanical analysis of surgeon's gesture in a laparoscopic virtual scenario. , 2006, Studies in health technology and informatics.

[13]  Peter I. Corke,et al.  A robotics toolbox for MATLAB , 1996, IEEE Robotics Autom. Mag..

[14]  Sylvie Gibet,et al.  Analysis of Human Motion, Based on the Reduction of Multidimensional Captured Data - Application to Hand Gesture Compression, Segmentation and Synthesis , 2008, AMDO.

[15]  Jakub Konecný,et al.  One-shot-learning gesture recognition using HOG-HOF features , 2014, J. Mach. Learn. Res..

[16]  Mehrdad Yazdani,et al.  A simple control policy for achieving minimum jerk trajectories , 2012, Neural Networks.

[17]  Ankit Chaudhary,et al.  Intelligent Approaches to interact with Machines using Hand Gesture Recognition in Natural way: A Survey , 2011, ArXiv.

[18]  Ayoub Al-Hamadi,et al.  Robust methods for hand gesture spotting and recognition using Hidden Markov Models and Conditional Random Fields , 2010, The 10th IEEE International Symposium on Signal Processing and Information Technology.

[19]  Ju Yong Chang Nonparametric Feature Matching Based Conditional Random Fields for Gesture Recognition from Multi-Modal Video , 2016, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[20]  Daan Wierstra,et al.  One-shot Learning with Memory-Augmented Neural Networks , 2016, ArXiv.

[21]  Terence Sim,et al.  The CMU Pose, Illumination, and Expression (PIE) database , 2002, Proceedings of Fifth IEEE International Conference on Automatic Face Gesture Recognition.

[22]  Ming Ouhyoung,et al.  A real-time continuous gesture recognition system for sign language , 1998, Proceedings Third IEEE International Conference on Automatic Face and Gesture Recognition.

[23]  Elisabetta Bevacqua,et al.  Real-time gesture recognition based on motion quality analysis , 2015, 2015 7th International Conference on Intelligent Technologies for Interactive Entertainment (INTETAIN).

[24]  Dan Xu,et al.  Online Dynamic Gesture Recognition for Human Robot Interaction , 2015, J. Intell. Robotic Syst..

[25]  Ching-Hua Chiu,et al.  Minimum Energy Expenditure of Arm and Leg Motions , 2009 .

[26]  Juan Pablo Wachs,et al.  A Human-Centered Approach to One-Shot Gesture Learning , 2017, Front. Robot. AI.

[27]  Kaveh Heidary,et al.  Exploring margin setting for good generalization in multiple class discrimination , 2005, Pattern Recognit..

[28]  Karthik Ramani,et al.  GestureAnalyzer: visual analytics for pattern analysis of mid-air hand gestures , 2014, SUI.

[29]  Nasser Kehtarnavaz,et al.  Real-time robust vision-based hand gesture recognition using stereo images , 2013, Journal of Real-Time Image Processing.

[30]  Peter I. Corke,et al.  A Simple and Systematic Approach to Assigning Denavit–Hartenberg Parameters , 2007, IEEE Transactions on Robotics.

[31]  Helena M. Mentis,et al.  Instructing people for training gestural interactive systems , 2012, CHI.

[32]  John Rasmussen,et al.  Modeling of Human Arm Energy Expenditure for Predicting Energy Optimal Trajectories , 2011 .

[33]  Ling Shao,et al.  One shot learning gesture recognition from RGBD images , 2012, 2012 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops.

[34]  Sotaro Kita,et al.  Movement Phase in Signs and Co-Speech Gestures, and Their Transcriptions by Human Coders , 1997, Gesture Workshop.

[35]  Jun Wan,et al.  Explore Efficient Local Features from RGB-D Data for One-Shot Learning Gesture Recognition , 2016, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[36]  Joshua B. Tenenbaum,et al.  Human-level concept learning through probabilistic program induction , 2015, Science.

[37]  Jeremy D Wong,et al.  The cost of moving optimally: kinematic path selection. , 2014, Journal of neurophysiology.

[38]  Juan Pablo Wachs,et al.  User-Centered and Analytic-Based Approaches to Generate Usable Gestures for Individuals With Quadriplegia , 2016, IEEE Transactions on Human-Machine Systems.