A Spatiotemporal Robust Approach for Human Activity Recognition

Nowadays, human activity recognition is considered to be one of the fundamental topics in computer vision research areas, including human-robot interaction. In this work, a novel method is proposed utilizing the depth and optical flow motion information of human silhouettes from video for human activity recognition. The recognition method utilizes enhanced independent component analysis (EICA) on depth silhouettes, optical flow motion features, and hidden Markov models (HMMs) for recognition. The local features are extracted from the collection of the depth silhouettes exhibiting various human activities. Optical flow-based motion features are also extracted from the depth silhouette area and used in an augmented form to form the spatiotemporal features. Next, the augmented features are enhanced by generalized discriminant analysis (GDA) for better activity representation. These features are then fed into HMMs to model human activities and recognize them. The experimental results show the superiority of the proposed approach over the conventional ones.

[1]  David Zhang,et al.  Human gait recognition by the fusion of motion and static spatio-temporal templates , 2007, Pattern Recognit..

[2]  Rama Chellappa,et al.  Identification of humans using gait , 2004, IEEE Transactions on Image Processing.

[3]  Toru Nakata Recognizing Human Activities in Video by Multi-resolutional Optical Flows , 2006, 2006 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[4]  Md. Zia Uddin,et al.  Independent shape component-based human activity recognition via Hidden Markov Model , 2010, Applied Intelligence.

[5]  Ankur Agarwal,et al.  Recovering 3D human pose from monocular images , 2006, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[6]  Jeung-Chul Park,et al.  3D Mesh Construction from Depth Images with Occlusion , 2006, PCM.

[7]  Antonio Fernández-Caballero,et al.  A survey of video datasets for human action and activity recognition , 2013, Comput. Vis. Image Underst..

[8]  Xia Liu,et al.  Hand gesture recognition using depth data , 2004, Sixth IEEE International Conference on Automatic Face and Gesture Recognition, 2004. Proceedings..

[9]  Nikolaos Doulamis,et al.  Local Geometrically Enriched Mixtures for Stable and Robust Human Tracking in Detecting Falls , 2013 .

[10]  Suki Kim,et al.  A Portable Surveillance Camera Architecture using One-bit Motion Detection , 2007, IEEE Transactions on Consumer Electronics.

[11]  Nicholas R. Howe,et al.  Silhouette lookup for monocular 3D pose tracking , 2007, Image Vis. Comput..

[12]  Munsang Kim,et al.  Intention reading from a fuzzy-based human engagement model and behavioural features , 2012 .

[13]  G. Iddan,et al.  3D IMAGING IN THE STUDIO (AND ELSEWHERE...) , 2001 .

[14]  Antonio Fernández-Caballero,et al.  Human activity monitoring by local and global finite state machines , 2012, Expert Syst. Appl..

[15]  Feng Niu,et al.  HMM-Based Segmentation and Recognition of Human Activities from Video Sequences , 2005, 2005 IEEE International Conference on Multimedia and Expo.

[16]  LinLin Shen,et al.  Gabor wavelets and General Discriminant Analysis for face identification and verification , 2007, Image Vis. Comput..

[17]  B. S. Manjunath,et al.  Probabilistic motion parameter models for human activity recognition , 2002, Object recognition supported by user interaction for service robots.

[18]  Francisco José Madrid-Cuevas,et al.  Depth silhouettes for gesture recognition , 2008, Pattern Recognit. Lett..

[19]  Lawrence R. Rabiner,et al.  A tutorial on hidden Markov models and selected applications in speech recognition , 1989, Proc. IEEE.

[20]  Dan Xu,et al.  Comparison of PCA, LDA and GDA for palmprint verification , 2010, 2010 International Conference on Information, Networking and Automation (ICINA).

[21]  David J. Fleet,et al.  Performance of optical flow techniques , 1994, International Journal of Computer Vision.