Recognition of human-human interaction using CWDTW

Understanding the activities of human is a challenging task in Computer Vision. Identifying the activities of human from videos and predicting their activity class label is the key functionality of Human Activity Recognition system. In general the major issues of Human activity recognition system is to identify the activities with or without the concurrent movement of body parts, occlusion, incremental learning etc. Among these issues the major difficulty lies in detecting the activity of human performing the interactions with or without the concurrent movement of their body parts. This paper aims in resolving the aforementioned problem. Here, the frames are extracted from the videos using the conventional frame extraction techniques. A pixel-based Local Binary Similarity Pattern background subtraction algorithm is used to detect the foreground from the extracted frames. The features are extracted from the detected foreground using Histogram of oriented Gradients and pyramidal feature extraction technique. A 20-point Microsoft human kinematic model is constructed using the set of features present in the frame and supervised temporal-stochastic neighbor embedding is applied to transform a high dimensional data to a low dimensional data. K-means clustering is then applied to produce a bag of key poses. The classifier Constrained Weighted Dynamic Time Warping(CWDTW) is used for the final generation of activity class label. Experimental results show the higher recognition rate achieved for various interactions with the benchmarking datasets such as Kinect Interaction dataset and Gaming dataset.

[1]  Guillaume-Alexandre Bilodeau,et al.  Improving background subtraction using Local Binary Similarity Patterns , 2014, IEEE Winter Conference on Applications of Computer Vision.

[2]  S. Chitrakala,et al.  Robust and Adaptive Approach for Human Action Recognition Based on Weighted Enhanced Dynamic Time Warping , 2015, WCI '15.

[3]  Alexandros André Chaaraoui,et al.  Optimizing human action recognition based on a cooperative coevolutionary algorithm , 2014, Eng. Appl. Artif. Intell..

[4]  Alexandros André Chaaraoui,et al.  Silhouette-based human action recognition using sequences of key poses , 2013, Pattern Recognit. Lett..

[5]  Hongsheng Li,et al.  Silhouette Analysis for Human Action Recognition Based on Supervised Temporal t-SNE and Incremental Learning , 2015, IEEE Transactions on Image Processing.

[6]  Larry S. Davis,et al.  Recognizing actions by shape-motion prototype trees , 2009, 2009 IEEE 12th International Conference on Computer Vision.

[7]  Mubarak Shah,et al.  Human Action Recognition in Videos Using Kinematic Features and Multiple Instance Learning , 2010, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[8]  Dimitrios Makris,et al.  G3Di: A Gaming Interaction Dataset with a Real Time Detection and Evaluation Framework , 2014, ECCV Workshops.

[9]  Ling Shao,et al.  Learning Discriminative Key Poses for Action Recognition , 2013, IEEE Transactions on Cybernetics.

[10]  Rama Chellappa,et al.  Human Action Recognition by Representing 3D Skeletons as Points in a Lie Group , 2014, 2014 IEEE Conference on Computer Vision and Pattern Recognition.

[11]  Dimitris Samaras,et al.  Two-person interaction detection using body-pose features and multiple instance learning , 2012, 2012 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops.

[12]  Marc Van Droogenbroeck,et al.  ViBe: A Universal Background Subtraction Algorithm for Video Sequences , 2011, IEEE Transactions on Image Processing.

[13]  Alexandros André Chaaraoui,et al.  Adaptive Human Action Recognition With an Evolving Bag of Key Poses , 2014, IEEE Transactions on Autonomous Mental Development.