Human motion understanding based on motion capture (mocap) data is investigated. Recent rapid developments and applications of mocap systems have resulted in a large corpus of mocap sequences, and an automated annotation technique that can classify basic motion types into multiple categories is needed. A novel technique for automated mocap data classification is developed in this work. Specifically, we adopt the tree-structured vector quantization (TSVQ) method to approximate human poses by codewords and approximate the dynamics of mocap sequences by a codeword sequence. To classify mocap data into different categories, we consider three approaches: 1) the spatial domain approach based on the histogram of codewords, 2) the spatial-time domain approach via codeword sequence matching, and 3) a decision fusion approach. We test the proposed algorithm on the CMU mocap database using the n-fold cross validation procedure and obtain a correct classification rate of 97%.
[1]
Peter Sanders,et al.
Simple Linear Work Suffix Array Construction
,
2003,
ICALP.
[2]
Michelle Effros.
Practical multi-resolution source coding: TSVQ revisited
,
1998,
Proceedings DCC '98 Data Compression Conference (Cat. No.98TB100225).
[3]
Hiroki Arimura,et al.
Linear-Time Longest-Common-Prefix Computation in Suffix Arrays and Its Applications
,
2001,
CPM.
[4]
Michelle Effros.
Practical Multi-Resolution Source Coding: TSVQ
,
1998,
DCC 1998.
[5]
Enno Ohlebusch,et al.
Replacing suffix trees with enhanced suffix arrays
,
2004,
J. Discrete Algorithms.
[6]
Nasser M. Nasrabadi,et al.
Image coding using vector quantization: a review
,
1988,
IEEE Trans. Commun..