GMM-QNT hybrid framework for vision-based human motion analysis

The understanding of human behaviour in video is a challenging task in that the same behaviour might have several different meanings depending upon the scene and task context in which it is performed. While human seem to perform scene interpretations without effort, this is a formidable and yet unsolved task for artificial vision systems. One of the main reasons is that there exists a gap between low-level vision at signal level and high-level representation of activities at symbolic level. In this paper, we present an intelligent connection framework using Gaussian Mixture Model-based clustering (GMM) to bridge the low-level vision data and the Qualitative Normalised Templates (QNT) - a symbolic representation for human motion based on fuzzy qualitative robot kinematics, which could link the former with domain-dependent scenarios. The proposed method has been applied to the recognition of eight types of human motions and an empirical comparison with fuzzy hidden Markov-based human motion recognition system.

[1]  Honghai Liu,et al.  Fuzzy Qualitative Robot Kinematics , 2008, IEEE Transactions on Fuzzy Systems.

[2]  Aaron F. Bobick,et al.  Recognition of Visual Activities and Interactions by Stochastic Parsing , 2000, IEEE Trans. Pattern Anal. Mach. Intell..

[3]  Honghai Liu,et al.  Fuzzy Qualitative Trigonometry , 2005, 2005 IEEE International Conference on Systems, Man and Cybernetics.

[4]  Aude Billard,et al.  On Learning, Representing, and Generalizing a Task in a Humanoid Robot , 2007, IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics).

[5]  Mubarak Shah,et al.  Monitoring human behavior from video taken in an office environment , 2001, Image Vis. Comput..

[6]  Changshui Zhang,et al.  Active curve axis Gaussian mixture models , 2005, Pattern Recognit..

[7]  Honghai Liu,et al.  A fuzzy qualitative approach to human motion recognition , 2008, 2008 IEEE International Conference on Fuzzy Systems (IEEE World Congress on Computational Intelligence).

[8]  Aaron F. Bobick,et al.  Parametric Hidden Markov Models for Gesture Recognition , 1999, IEEE Trans. Pattern Anal. Mach. Intell..

[9]  Alex Pentland,et al.  Coupled hidden Markov models for complex action recognition , 1997, Proceedings of IEEE Computer Society Conference on Computer Vision and Pattern Recognition.

[10]  Barbara Caputo,et al.  Recognizing human actions: a local SVM approach , 2004, Proceedings of the 17th International Conference on Pattern Recognition, 2004. ICPR 2004..

[11]  Michael Isard,et al.  CONDENSATION—Conditional Density Propagation for Visual Tracking , 1998, International Journal of Computer Vision.

[12]  Paul D. Gader,et al.  Generalized hidden Markov models. I. Theoretical frameworks , 2000, IEEE Trans. Fuzzy Syst..

[13]  Ronen Basri,et al.  Actions as space-time shapes , 2005, Tenth IEEE International Conference on Computer Vision (ICCV'05) Volume 1.

[14]  Paul D. Gader,et al.  Generalized hidden Markov models. II. Application to handwritten word recognition , 2000, IEEE Trans. Fuzzy Syst..

[15]  Alex Pentland,et al.  A Bayesian Computer Vision System for Modeling Human Interactions , 1999, IEEE Trans. Pattern Anal. Mach. Intell..

[16]  Xavier Varona,et al.  aSpaces : Action Spaces for Recognition and Synthesis of Human Actions , 2002, AMDO.

[17]  D. Kendall A Survey of the Statistical Theory of Shape , 1989 .

[18]  Rama Chellappa,et al.  Activity recognition using the dynamics of the configuration of interacting objects , 2003, 2003 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2003. Proceedings..