Subunit Boundary Detection for Sign Language Recognition Using Spatio-temporal Modelling

The use of subunits offers a feasible way to recognize sign language with large vocabulary. The initial step is to partition signs into elementary units. In this paper, we firstly define a subunit as one continuous hand action in time and space, which comprises a series of interrelated consecutive frames. Then, we propose a solution to detect the subunit boundary according to spatiotemporal features using a three-stage hierarchy: in the first stage, we apply hand segmentation and tracking algorithm to capture motion speeds and trajectories; in the second stage, the obtained speed and trajectory information are combined to locate subunit boundaries; finally, temporal clustering by dynamic time warping (DTW) is adopted to merge similar segments and refine the results. The presented work does not need prior knowledge of the types of signs and is robust to signer behaviour variation. Moreover, it can provide a base for highlevel sign language understanding. Experiments on many real-world signing videos show the effectiveness of the proposed work.

[1]  Scott K. Liddell,et al.  American Sign Language: The Phonological Base , 2013 .

[2]  Alex Pentland,et al.  Real-Time American Sign Language Recognition Using Desk and Wearable Computer Based Video , 1998, IEEE Trans. Pattern Anal. Mach. Intell..

[3]  George Awad,et al.  A Unified System for Segmentation and Tracking of Face and Hands in Sign Language Recognition , 2006, 18th International Conference on Pattern Recognition (ICPR'06).

[4]  George Awad,et al.  Automatic Skin Segmentation for Gesture Recognition Combining Region and Support Vector Machine Active Learning , 2006, 7th International Conference on Automatic Face and Gesture Recognition (FGR06).

[5]  Dimitris N. Metaxas,et al.  Toward Scalability in ASL Recognition: Breaking Down Signs into Phonemes , 1999, Gesture Workshop.

[6]  Mohammed Yeasin,et al.  Visual understanding of dynamic hand gestures , 2000, Pattern Recognit..

[7]  Narendra Ahuja,et al.  Extraction of 2D Motion Trajectories and Its Application to Hand Gesture Recognition , 2002, IEEE Trans. Pattern Anal. Mach. Intell..

[8]  Ming Ouhyoung,et al.  A real-time continuous gesture recognition system for sign language , 1998, Proceedings Third IEEE International Conference on Automatic Face and Gesture Recognition.

[9]  Yangsheng Xu,et al.  Trajectory fitting with smoothing splines using velocity information , 2000, Proceedings 2000 ICRA. Millennium Conference. IEEE International Conference on Robotics and Automation. Symposia Proceedings (Cat. No.00CH37065).

[10]  Nobuyuki Otsu,et al.  ATlreshold Selection Method fromGray-Level Histograms , 1979 .

[11]  Karl-Friedrich Kraiss,et al.  Towards an Automatic Sign Language Recognition System Using Subunits , 2001, Gesture Workshop.

[12]  Wen Gao,et al.  A novel approach to automatically extracting basic units from Chinese sign language , 2004, Proceedings of the 17th International Conference on Pattern Recognition, 2004. ICPR 2004..

[13]  A E Marble,et al.  Image processing system for interpreting motion in American Sign Language. , 1992, Journal of biomedical engineering.

[14]  W. Stokoe,et al.  Sign language structure: an outline of the visual communication systems of the American deaf. 1960. , 1961, Journal of deaf studies and deaf education.