Semantic Matchmaking for Kinect-Based Posture and Gesture Recognition

Innovative analysis methods applied to data extracted by off-the-shelf peripherals can provide useful results in activity recognition without requiring large computational resources. In this paper a framework is proposed for automated posture and gesture recognition, exploiting depth data provided by a commercial tracking device. The detection problem is handled as a semantic-based resource discovery. A general data model and the corresponding ontology provide the formal underpinning for automatic posture and gesture annotation via standard Semantic Web languages. Hence, a logic-based matchmaking, exploiting non-standard inference services, allows to: (i) detect postures via on-the-fly comparison of the retrieved annotations with standard posture descriptions stored as instances of a proper Knowledge Base, (ii) compare subsequent postures in order to recognize gestures. The framework has been implemented in a prototypical tool and experimental tests have been carried out on a reference dataset. Preliminary results indicate the feasibility of the proposed approach.

[1]  Diego Calvanese,et al.  The Description Logic Handbook , 2007 .

[2]  Tommaso Di Noia,et al.  A Nonmonotonic Approach to Semantic Matchmaking and Request Refinement in E-Marketplaces , 2007, Int. J. Electron. Commer..

[3]  F. Xavier Roca,et al.  Semantics of Human Behavior in Image Sequences , 2011, Computer Analysis of Human Behavior.

[4]  Xiang Cao,et al.  DejaVu: integrated support for developing interactive camera-based programs , 2012, UIST.

[5]  João Marcelo X. N. Teixeira,et al.  Increasing kinect application development productivity by an enhanced hardware abstraction , 2012, EICS '12.

[6]  Debasis Mitra,et al.  Qualitative Direction Calculi with Arbitrary Granularity , 2004, PRICAI.

[7]  Daqing Zhang,et al.  Gesture Recognition with a 3-D Accelerometer , 2009, UIC.

[8]  Ramakant Nevatia,et al.  VERL: An Ontology Framework for Representing and Annotating Video Events , 2005, IEEE Multim..

[9]  Kanad K. Biswas,et al.  Gesture recognition using Microsoft Kinect® , 2011, The 5th International Conference on Automation, Robotics and Applications.

[10]  Vassilis Athitsos,et al.  Comparing gesture recognition accuracy using color and depth information , 2011, PETRA '11.

[11]  Álvaro García-Martín,et al.  An Ontology for Event Detection and its Application in Surveillance Video , 2009, AVSS.

[12]  Darko Kirovski,et al.  Real-time classification of dance gestures from skeleton animation , 2011, SCA '11.

[13]  Miguel A. Patricio,et al.  Ontology-based context representation and reasoning for object tracking and scene interpretation in video , 2011, Expert Syst. Appl..

[14]  Helena M. Mentis,et al.  Instructing people for training gestural interactive systems , 2012, CHI.

[15]  Eugenio Di Sciascio,et al.  Concept abduction and contraction in semantic-based P2P environments , 2011, Web Intell. Agent Syst..

[16]  Saïd Mahmoudi,et al.  An Ontology for video human movement representation based on Benesh notation , 2012, 2012 International Conference on Multimedia Computing and Systems.

[17]  Eugenio Di Sciascio,et al.  Mini-ME: the Mini Matchmaking Engine , 2012, ORE.

[18]  Bernt Schiele,et al.  Unsupervised Discovery of Structure in Activity Data Using Multiple Eigenspaces , 2006, LoCA.

[19]  Hector J. Levesque,et al.  The Tractability of Subsumption in Frame-Based Description Languages , 1984, AAAI.

[20]  Hao Zhang,et al.  Kinect Gesture Recognition for Interactive System , 2012 .

[21]  Chris D. Nugent,et al.  Ontology-based activity recognition in intelligent pervasive environments , 2009, Int. J. Web Inf. Syst..

[22]  Mario Fernando Montenegro Campos,et al.  Real-Time Gesture Recognition from Depth Data through Key Poses Learning and Decision Forests , 2012, 2012 25th SIBGRAPI Conference on Graphics, Patterns and Images.

[23]  Changsheng Xu,et al.  A Generic Framework for Video Annotation via Semi-Supervised Learning , 2012, IEEE Transactions on Multimedia.