Video-based feature extraction techniques for isolated arabic sign language recognition

This paper presents various spatio-temporal feature extraction techniques with applications to recognition of isolated Arabic sign language (ArSL) gestures. The temporal features of a video-based gesture are extracted through forward image predictions. The prediction errors are thresholded and accumulated into one image that represents the sequence motion. The motion representation is then followed by spatial domain feature extractions, namely; 2-D DCT followed by zonal coding or Radon transformation followed by ideal low pass filtering of the projected spatial features. The proposed feature extraction scheme was complemented by simple classification techniques, namely, KNN and Bayesian classifiers. Experimental results showed superior classification performance ranging from 97% to 100% recognition rates. To validate our proposed technique, we conducted a series of experiments using the classical way of classifying data with temporal dependencies. Namely, hidden Markov models (HMMs). Here, the features are the consecutive binarized image differences, each of which is followed by spatial domain feature extraction schemes. Experimental results revealed that the proposed feature extraction scheme combined with simple KNN or Bayesian classification yields comparable results to the classical HMM-based scheme.

[1]  Alex Pentland,et al.  Real-Time American Sign Language Recognition Using Desk and Wearable Computer Based Video , 1998, IEEE Trans. Pattern Anal. Mach. Intell..

[2]  Yuntao Cui,et al.  Appearance-Based Hand Sign Recognition from Intensity Image Sequences , 2000, Comput. Vis. Image Underst..

[3]  Khaled Assaleh,et al.  Recognition of Arabic Sign Language Alphabet Using Polynomial Classifiers , 2005, EURASIP J. Adv. Signal Process..

[4]  Omar M. Al-Jarrah,et al.  Recognition of gestures in Arabic sign language using neuro-fuzzy systems , 2001, Artif. Intell..

[5]  J. Todd Book Review: Digital image processing (second edition). By R. C. Gonzalez and P. Wintz, Addison-Wesley, 1987. 503 pp. Price: £29.95. (ISBN 0-201-11026-1) , 1988 .

[6]  Z. Zenn Bien,et al.  A dynamic gesture recognition system for the Korean sign language (KSL) , 1996, IEEE Trans. Syst. Man Cybern. Part B.

[7]  Karl-Friedrich Kraiss,et al.  Video-based sign recognition using self-organizing subunits , 2002, Object recognition supported by user interaction for service robots.

[8]  Wen Gao,et al.  A Real-Time Large Vocabulary Continuous Recognition System for Chinese Sign Language , 2001, IEEE Pacific Rim Conference on Multimedia.

[9]  Shan Lu,et al.  The Recognition Algorithm with Non-contact for Japanese Sign Language Using Morphological Analysis , 1997, Gesture Workshop.

[10]  Rafael C. González,et al.  Local Determination of a Moving Contrast Edge , 1985, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[11]  Mohammed Waleed Kadous,et al.  Machine Recognition of Auslan Signs Using PowerGloves: Towards Large-Lexicon Recognition of Sign Lan , 1996 .

[12]  Surendra Ranganath,et al.  Automatic Sign Language Analysis: A Survey and the Future beyond Lexical Meaning , 2005, IEEE Trans. Pattern Anal. Mach. Intell..