Human-inspired search for redundancy in automatic sign language recognition

Human perception of sign language can serve as inspiration for the improvement of automatic recognition systems. Experiments with human signers show that sign language signs contain redundancy over time. In this article, experiments are conducted to investigate whether comparable redundancies also exist for an automatic sign language recognition system. Such redundancies could be exploited, for example, by reserving more processing resources for the more informative phases of a sign, or by discarding uninformative phases. In the experiments, an automatic system is trained and tested on isolated fragments of sign language signs. The stimuli used were similar to those of the human signer experiments, allowing us to compare the results. The experiments show that redundancy over time exists for the automatic recognizer. The central phase of a sign is the most informative phase, and the first half of a sign is sufficient to achieve a recognition performance similar to that of the entire sign. These findings concur with the results of the human signer studies. However, there are differences as well, most notably the fact that human signers score better on the early phases of a sign than the automatic system. The results can be used to improve the automatic recognizer, by using only the most informative phases of a sign as input.

[1]  Liya Ding,et al.  Modelling and recognition of the linguistic components in American Sign Language , 2009, Image Vis. Comput..

[2]  D. McNeill Hand and Mind , 1995 .

[3]  Alex Pentland,et al.  Invariant features for 3-D gesture recognition , 1996, Proceedings of the Second International Conference on Automatic Face and Gesture Recognition.

[4]  Marcel J. T. Reinders,et al.  Person-Independent 3D Sign Language Recognition , 2007, Gesture Workshop.

[5]  Marcel J. T. Reinders,et al.  Sign language perception research for improving automatic sign language recognition , 2009, Electronic Imaging.

[6]  Andrea J. van Doorn,et al.  When do people start to recognize signs , 2009 .

[7]  Andrea J. van Doorn,et al.  When and how well do people see the onset of gestures , 2007 .

[8]  Dimitris N. Metaxas,et al.  Handshapes and Movements: Multiple-Channel American Sign Language Recognition , 2003, Gesture Workshop.

[9]  F. Grosjean,et al.  Sign Recognition Processes in American Sign Language: the Effect of Context , 1982 .

[10]  Karl-Friedrich Kraiss,et al.  Rapid signer adaptation for continuous sign language recognition using a combined approach of eigenvoices, MLLR, and MAP , 2008, 2008 19th International Conference on Pattern Recognition.

[11]  George Awad,et al.  Modelling and segmenting subunits for sign language recognition based on hand motion analysis , 2009, Pattern Recognit. Lett..

[12]  K. Emmorey Language, Cognition, and the Brain: Insights From Sign Language Research , 2001 .

[13]  David Windridge,et al.  A Linguistic Feature Vector for the Visual Interpretation of Sign Language , 2004, ECCV.

[14]  Joseph B. Kruskall,et al.  The Symmetric Time-Warping Problem : From Continuous to Discrete , 1983 .

[15]  Wendy Sandler,et al.  Sign Language and Linguistic Universals: Entering the lexicon: lexicalization, backformation, and cross-modal borrowing , 2006 .

[16]  Karl-Friedrich Kraiss,et al.  Robust Person-Independent Visual Sign Language Recognition , 2005, IbPRIA.

[17]  M. B. Waldron,et al.  Isolated ASL sign recognition system for deaf persons , 1995 .

[18]  F. Grosjean Sign & Word Recognition: A First Comparison , 2013 .

[19]  Diane Brentari,et al.  A Prosodic Model of Sign Language Phonology , 1999 .

[20]  K. Emmorey,et al.  Lexical Recognition in Sign Language: Effects of Phonetic Structure and Morphology , 1990, Perceptual and motor skills.

[21]  W. Stokoe,et al.  Sign language structure: an outline of the visual communication systems of the American deaf. 1960. , 1961, Journal of deaf studies and deaf education.

[22]  E. A. Hendriks,et al.  Which Fragments of a Sign Enable Its Recognition? , 2009 .

[23]  John K. Tsotsos,et al.  Definition and recovery of kinematic features for recognition of American sign language movements , 2008, Image Vis. Comput..

[24]  Marcel J. T. Reinders,et al.  Sign Language Recognition by Combining Statistical DTW and Independent Classification , 2008, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[25]  Alex Pentland,et al.  Real-Time American Sign Language Recognition Using Desk and Wearable Computer Based Video , 1998, IEEE Trans. Pattern Anal. Mach. Intell..

[26]  Robert P. W. Duin,et al.  Using two-class classifiers for multiclass classification , 2002, Object recognition supported by user interaction for service robots.

[27]  Jeroen Lichtenauer,et al.  3 D versus 2 D Pose Information for Recognition of NGT Signs , 2006 .

[28]  Sotaro Kita,et al.  Movement Phase in Signs and Co-Speech Gestures, and Their Transcriptions by Human Coders , 1997, Gesture Workshop.

[29]  Marcel J. T. Reinders,et al.  Influence of Handshape Information on Automatic Sign Language Recognition , 2009, Gesture Workshop.

[30]  Justine Cassell,et al.  Recovering the temporal structure of natural gesture , 1996, Proceedings of the Second International Conference on Automatic Face and Gesture Recognition.

[31]  W. Stokoe,et al.  A dictionary of American sign language on linguistic principles , 1965 .

[32]  David G. Stork,et al.  Pattern Classification (2nd ed.) , 1999 .