Collecting and evaluating the CUNY ASL corpus for research on American Sign Language animation

Automated synthesis of American Sign Language (ASL) animations will benefit people who are deaf with low English literacy.Annotated ASL motion-capture corpora enable researchers to produce animations with complex spatial and linguistic phenomena.We built the first large, annotated, motion-capture, unscripted ASL corpus, which enables new animation synthesis research.We evaluated our corpus quality in several ways, including user-based animation-comprehension studies by native ASL signers. While there is great potential for sign language animation generation software to improve the accessibility of information for deaf individuals with low written-language literacy, the understandability of current sign language animation systems is limited. Data-driven methodologies using annotated sign language corpora encoding detailed human movement have enabled some researchers to address several key linguistic challenges in ASL generation. This article motivates and describes our current research on collecting a motion-capture corpus of American Sign Language (ASL). As an evaluation of our motion-capture configuration, calibration, and recording protocol, we have conducted several rounds of evaluation studies with native ASL signers, and we have made use of our collected data to synthesize novel animations of ASL, which have also been evaluated in experimental studies with native signers.

[1]  Kostas Karpouzis,et al.  A knowledge-based sign synthesis architecture , 2008, Universal Access in the Information Society.

[2]  Peter Vamplew Recognition of sign language gestures using neural networks , 1996 .

[3]  Mitchell P. Marcus,et al.  Generating american sign language classifier predicates for english-to-asl machine translation , 2006 .

[4]  Constantine Stephanidis,et al.  Universal access in the information society , 1999, HCI.

[5]  Matt Huenerfauth,et al.  Modeling and synthesizing spatially inflected verbs for American sign language animations , 2010, ASSETS '10.

[6]  Eric Auer,et al.  Combining video and numeric data in the analysis of sign languages with the ELAN annotation software , 2006 .

[7]  Éva Sáfár,et al.  The Architecture of an English-Text-to-Sign-Languages Translation System , 2001 .

[8]  Kearsy Cormier,et al.  Grammaticization of indexic signs : how American Sign Language expresses numerosity , 2002 .

[9]  Zeungnam Bien,et al.  Real-time recognition system of Korean sign language based on elementary components , 1997, Proceedings of 6th International Fuzzy Systems Conference.

[10]  Pengfei Lu,et al.  Synthesizing American Sign Language Spatially Inflected Verbs from Motion-Capture Data , 2011 .

[11]  Matt Huenerfauth,et al.  Data-Driven Synthesis of Spatially Inflected Verbs for American Sign Language Animation , 2011, TACC.

[12]  Abdul Nasser S. Abu-Rezq,et al.  Arabic glove-talk (AGT): A communication aid for vocally impaired , 1998, Pattern Analysis and Applications.

[13]  Hideki Hirakawa,et al.  A machine translation system and method , 1987 .

[14]  Norman I. Badler,et al.  A machine translation system from English to American Sign Language , 2000, AMTA.

[15]  Matt Huenerfauth,et al.  A Linguistically Motivated Model for Speed and Pausing in Animations of American Sign Language , 2009, TACC.

[16]  Andy Way,et al.  An Example-Based Approach to Translating Sign Language , 2005, MTSUMMIT.

[17]  Onno Crasborn,et al.  Workshop Proceedings: 5th Workshop on the Representation and Processing of Sign Languages: Interactions between Corpus and Lexicon , 2012, LREC 2012.

[18]  Andrew Rosenberg,et al.  Evaluating importance of facial expression in american sign language and pidgin signed english animations , 2011, ASSETS.

[19]  Daan Broeder,et al.  Sharing sign language corpora online: proposals for transcription and metadata categories , 2004 .

[20]  Susan Lioyd McBurney,et al.  Modality and structure in signed and spoken languages: Pronominal reference in signed and spoken language: Are grammatical categories modality-dependent? , 2002 .

[21]  Tony Veale,et al.  The Challenges of Cross-Modal Translation: English-to-Sign-Language Translation in the Zardoz System , 1998, Machine Translation.

[22]  Kearsy Cormier,et al.  Modality and structure in signed and spoken languages: Frontmatter , 2002 .

[23]  Matt Huenerfauth,et al.  Collecting a Motion-Capture Corpus of American Sign Language for Data-Driven Generation Research , 2010, SLPAT@NAACL.

[24]  Matt Huenerfauth,et al.  Accessible motion-capture glove calibration protocol for recording sign language data from deaf subjects , 2009, Assets '09.

[25]  Ian Marshall,et al.  Linguistic modelling and language-processing technologies for Avatar-based sign language presentation , 2008, Universal Access in the Information Society.

[26]  Mark Wells,et al.  Tessa, a system to aid communication with deaf people , 2002, ASSETS.

[27]  Stavroula-Evita Fotinea,et al.  GSLC: Creation and Annotation of a Greek Sign Language Corpus for HCI , 2007, HCI.

[28]  Hermann Ney,et al.  A German Sign Language Corpus of the Domain Weather Report , 2006, LREC.

[29]  Matt Huenerfauth,et al.  Evaluation of a psycholinguistically motivated timing model for animations of american sign language , 2008, Assets '08.

[30]  C. B. Traxler,et al.  The Stanford Achievement Test, 9th Edition: National Norming and Performance Standards for Deaf and Hard-of-Hearing Students. , 2000, Journal of deaf studies and deaf education.

[31]  John R. W. Glauert,et al.  Providing signed content on the Internet by synthesized animation , 2007, TCHI.

[32]  Sudeep Sarkar,et al.  Progress in Automated Computer Recognition of Sign Language , 2004, ICCHP.

[33]  Wen Gao,et al.  An approach based on phonemes to large vocabulary Chinese sign language recognition , 2002, Proceedings of Fifth IEEE International Conference on Automatic Face Gesture Recognition.

[34]  Matt Huenerfauth,et al.  Effect of spatial reference and verb inflection on the usability of sign language animations , 2011, Universal Access in the Information Society.

[35]  Brad Wieners,et al.  Reality check , 2018, Nature Energy.

[36]  Ian W. Marshall,et al.  Sign language generation using HPSG. , 2002, TMI.

[37]  C. Padden Interaction of Morphology and Syntax in American Sign Language , 2016 .

[38]  Matt Huenerfauth,et al.  Learning a Vector-Based Model of American Sign Language Inflecting Verbs from Motion-Capture Data , 2012, SLPAT@HLT-NAACL.

[39]  Annelies Braffort,et al.  Toward the Study of Sign Language Coarticulation: Methodology Proposal , 2009, 2009 Second International Conferences on Advances in Computer-Human Interactions.

[40]  Matt Huenerfauth Eliciting Spatial Reference for a Motion-Capture Corpus of American Sign Language Discourse , 2010 .

[41]  Nicolas Courty,et al.  The SignCom system for data-driven animation of interactive virtual signers , 2011, ACM Trans. Interact. Intell. Syst..

[42]  R. Mitchell,et al.  How Many People Use ASL in the United States? Why Estimates Need Updating , 2006 .

[43]  Scott K. Liddell Grammar, Gesture, and Meaning in American Sign Language , 2003 .

[44]  Stan Sclaroff,et al.  Challenges in development of the American Sign Language Lexicon Video Dataset (ASLLVD) corpus , 2012 .

[45]  Z. Zenn Bien,et al.  A dynamic gesture recognition system for the Korean sign language (KSL) , 1996, IEEE Trans. Syst. Man Cybern. Part B.

[46]  Pengfei Lu,et al.  Effect of presenting video as a baseline during an american sign language animation user study , 2012, ASSETS '12.