Toward the markerless and automatic analysis of kinematic features: A toolkit for gesture and movement research

Action, gesture, and sign represent unique aspects of human communication that use form and movement to convey meaning. Researchers typically use manual coding of video data to characterize naturalistic, meaningful movements at various levels of description, but the availability of markerless motion-tracking technology allows for quantification of the kinematic features of gestures or any meaningful human movement. We present a novel protocol for extracting a set of kinematic features from movements recorded with Microsoft Kinect. Our protocol captures spatial and temporal features, such as height, velocity, submovements/strokes, and holds. This approach is based on studies of communicative actions and gestures and attempts to capture features that are consistently implicated as important kinematic aspects of communication. We provide open-source code for the protocol, a description of how the features are calculated, a validation of these features as quantified by our protocol versus manual coders, and a discussion of how the protocol can be applied. The protocol effectively quantifies kinematic features that are important in the production (e.g., characterizing different contexts) as well as the comprehension (e.g., used by addressees to understand intent and semantics) of manual acts. The protocol can also be integrated with qualitative analysis, allowing fast and objective demarcation of movement units, providing accurate coding even of complex movements. This can be useful to clinicians, as well as to researchers studying multimodal communication or human–robot interactions. By making this protocol available, we hope to provide a tool that can be applied to understanding meaningful movement characteristics in human communication.

[1]  H. Bekkering,et al.  Communicative intent modulates production and comprehension of actions and gestures: A Kinect study , 2018, Cognition.

[2]  Martha W. Alibali,et al.  Why Gesture?: How the hands function in speaking, thinking and communicating , 2017 .

[3]  Leif Johannsen,et al.  Tool use without a tool: kinematic characteristics of pantomiming as compared to actual use and the effect of brain damage , 2012, Experimental Brain Research.

[4]  Günther Knoblich,et al.  Distinct Kinematic Markers of Demonstration and Joint Action Coordination? Evidence From Virtual Xylophone Playing , 2017, Journal of experimental psychology. Human perception and performance.

[5]  Georg Goldenberg,et al.  Defective pantomime of object use in left brain damage: apraxia or asymbolia? , 2003, Neuropsychologia.

[6]  Kade Paterson,et al.  Instrumenting gait assessment using the Kinect in people living with stroke: reliability and association with balance tests , 2015, Journal of NeuroEngineering and Rehabilitation.

[7]  Peter Wittenburg,et al.  ELAN: a Professional Framework for Multimodality Research , 2006, LREC.

[8]  Sotaro Kita,et al.  Movement Phase in Signs and Co-Speech Gestures, and Their Transcriptions by Human Coders , 1997, Gesture Workshop.

[9]  Kanad K. Biswas,et al.  Gesture recognition using Microsoft Kinect® , 2011, The 5th International Conference on Automation, Robotics and Applications.

[10]  Adso Fernández-Baena,et al.  Biomechanical Validation of Upper-Body and Lower-Body Joint Movements of Kinect Motion Capture Data for Rehabilitation Treatments , 2012, 2012 Fourth International Conference on Intelligent Networking and Collaborative Systems.

[11]  R A Abrams,et al.  Optimality in human motor performance: ideal control of rapid aimed movements. , 1988, Psychological review.

[12]  Janet Beavin Bavelas,et al.  Gesturing on the telephone: Independent effects of dialogue and visibility. , 2008 .

[13]  M. Gentilucci,et al.  Gaze Direction and Request Gesture in Social Interactions , 2012, PloS one.

[14]  P. Olivier,et al.  Accuracy of the Microsoft Kinect sensor for measuring movement in people with Parkinson's disease. , 2014, Gait & posture.

[15]  Nassir Navab,et al.  Motor Rehabilitation Using Kinect: A Systematic Review. , 2015, Games for health journal.

[16]  Didier Stricker,et al.  Comparison of Kinect V1 and V2 Depth Images in Terms of Accuracy and Precision , 2016, ACCV Workshops.

[17]  Peter Hagoort,et al.  Getting to the Point: The Influence of Communicative Intent on the Kinematics of Pointing Gestures , 2013, CogSci.

[18]  U. Castiello,et al.  Does the intention to communicate affect action kinematics? , 2009, Consciousness and Cognition.

[19]  Sotaro Kita,et al.  Attention to Speech-Accompanying Gestures: Eye Movements and Information Uptake , 2009, Journal of nonverbal behavior.

[20]  V. Manera,et al.  Cooperation or competition? Discriminating between social intentions by observing prehensile movements , 2011, Experimental Brain Research.

[21]  J. Holler,et al.  A third-person perspective on co-speech action gestures in Parkinson's disease , 2016, Cortex.

[22]  D. McNeill Hand and Mind: What Gestures Reveal about Thought , 1992 .

[23]  Stijn De Beugher,et al.  A semi-automatic annotation tool for unobtrusive gesture analysis , 2017, Language Resources and Evaluation.

[24]  Chris Rorden,et al.  Pantomime of tool use depends on integrity of left inferior frontal cortex. , 2007, Cerebral cortex.

[25]  Michael Kipp,et al.  ANVIL - a generic annotation tool for multimodal dialogue , 2001, INTERSPEECH.

[26]  S. Cook,et al.  Bridging gaps in common ground: Speakers design their gestures for their listeners. , 2016, Journal of experimental psychology. Learning, memory, and cognition.

[27]  J. Delafield-Butt,et al.  Toward the Autism Motor Signature: Gesture patterns during smart tablet gameplay identify children with autism , 2018 .

[28]  TeichriebVeronica,et al.  Motor Rehabilitation Using Kinect: A Systematic Review , 2015 .

[29]  supF.A. Elmisery,et al.  Motion Control of Robot by using Kinect Sensor , 2014 .

[30]  K M Heilman,et al.  Pantomime comprehension and ideomotor apraxia. , 1985, Journal of neurology, neurosurgery, and psychiatry.

[31]  Erkan Bostanci,et al.  Augmented reality applications for cultural heritage using Kinect , 2015, Human-centric Computing and Information Sciences.

[32]  K. Rohlfing,et al.  People modify their tutoring behavior in robot-directed interaction for action learning , 2009, 2009 IEEE 8th International Conference on Development and Learning.

[33]  Dare A. Baldwin,et al.  Evidence for ‘motionese’: modifications in mothers’ infant-directed action , 2002 .

[34]  Janet Beavin Bavelas,et al.  Linguistic influences on gesture’s form , 2005 .

[35]  Evaggelos Spyrou,et al.  A Real-Time Approach for Gesture Recognition using the Kinect Sensor , 2016, SETN.

[36]  Asli Ozyurek,et al.  Iconicity as a communicative strategy: Recipient design in multimodal demonstrations for adults and children , 2013 .