Generating american sign language classifier predicates for english-to-asl machine translation

A majority of deaf 18-year-olds in the United States have an English reading level below that of a typical 10-year-old student, and so machine translation (MT) software that could translate English text into American Sign Language (ASL) animations could significantly improve these individuals' access to information, communication, and services. Previous English-to-ASL MT projects have made limited progress by restricting their output to subsets of ASL phenomena---thus avoiding important linguistic and animation issues. None of these systems have shown how to generate classifier predicates (CPs), a phenomenon in which signers use special hand movements to indicate the location and movement of invisible objects (representing entities under discussion) in space around their bodies. CPs are frequent in ASL and are necessary for conveying many concepts. This project has created an English-to-ASL MT design capable of producing classifier predicates. The classifier predicate generator inside this design has a planning-based architecture that uses a 3D "visualization" model of the arrangement of objects in a scene discussed by the English input text. This generator would be one pathway in a multi-path English-to-ASL MT design; a separate processing pathway would be used to generate classifier predicates, to generate other ASL sentences, and to generate animations of Signed English (if the system lacked lexical resources for some input). Instead of representing the ASL animation as a string (of individual signs to perform), this system encodes the multimodal language signal as multiple channels that are hierarchically structured and coordinated over time. While this design feature and others have been prompted by the unique requirements of generating a sign language, these technologies have applications for the machine translation of written languages, the representation of other multimodal language signals, and the production of meaningful gestures by other animated virtual human characters. To evaluate the functionality and scalability of the most novel portion of this English-to-ASL MT design, this project has implemented a prototype-version of the planning-based classifier predicate generator. The classifier predicate animations produced by the system have been shown to native ASL signers to evaluate the output.

[1]  A. Mital,et al.  Effects of sitting and standing, reach distance, and arm orientation on isokinetic pull strengths in the horizontal plane , 1990 .

[2]  R. Elliott,et al.  An Overview of the SiGML Notation and SiGMLSigning Software System , 2004 .

[3]  Richard Sproat,et al.  WordsEye: an automatic text-to-scene conversion system , 2001, SIGGRAPH.

[4]  Ceil Lucas,et al.  Turn-taking mechanisms and active participation in meetings with Deaf and hearing participants in Flanders , 2002 .

[5]  Ian Marshall,et al.  Sign Language Translation via DRT and HPSG , 2002, CICLing.

[6]  Stefan Kopp,et al.  Towards integrated microplanning of language and iconic gesture for multimodal output , 2004, ICMI '04.

[7]  Norman I. Badler,et al.  Parameterized Action Representation and Natural Language Instructions for Dynamic Behavior Modification of Embodied Agents , 2000 .

[8]  Éva Sáfár,et al.  Extraction of Semantic Representations from Syntactic CMU Link Grammar linkages , 2001 .

[9]  Tony Veale,et al.  The Challenges of Cross-Modal Translation: English-to-Sign-Language Translation in the Zardoz System , 1998, Machine Translation.

[10]  Matt Huenerfauth American Sign Language Spatial Representations for an Accessible User-Interface , 2005 .

[11]  Marina L. McIntire,et al.  On the other hand : new perspectives on American sign language , 1979 .

[12]  Leonard Talmy The representation of spatial structure in spoken and signed language , 2004 .

[13]  Catherine N. Ball,et al.  Representation of american sign language for machine translation , 2002 .

[14]  Dana S. Nau,et al.  SHOP2: An HTN Planning System , 2003, J. Artif. Intell. Res..

[15]  Norman I. Badler,et al.  Visual Attention and Eye Gaze During Multiparty Conversations with Distractions , 2006, IVA.

[16]  Geoffrey Restall Coulter Current issues in ASL phonology , 1993 .

[17]  C Neidle,et al.  SignStream: A tool for linguistic and computer vision research on visual-gestural language data , 2001, Behavior research methods, instruments, & computers : a journal of the Psychonomic Society, Inc.

[18]  Norman I. Badler,et al.  Real-time reach planning for animated characters using hardware acceleration , 2003, Proceedings 11th IEEE International Workshop on Program Comprehension.

[19]  Norman I. Badler,et al.  Eyes alive , 2002, ACM Trans. Graph..

[20]  Stefan Kopp,et al.  Content in Context: Generating Language and Iconic Gestures without a Gestionary , 2004 .

[21]  Scott K. Liddell,et al.  American Sign Language: The Phonological Base , 2013 .

[22]  Pamela W. Jordan,et al.  A survey of current paradigms in machine translation , 1999, Adv. Comput..

[23]  Hao Yan,et al.  Coordination and context-dependence in the generation of embodied conversation , 2000, INLG.

[24]  Scott K. Liddell An investigation into the syntactic structure of American sign language , 1977 .

[25]  L. Stark,et al.  Most naturally occurring human saccades have magnitudes of 15 degrees or less. , 1975, Investigative ophthalmology.

[26]  D. MacLaughlin The structure of determiner phrases : evidence from American sign language , 1997 .

[27]  Nina Suszczańska TRANSLATING POLISH TEXTS INTO SIGN LANGUAGE IN THE TGT SYSTEM , 2002 .

[28]  Jill P. Morford,et al.  Frequency Characteristics of American Sign Language , 2003 .

[29]  Nancy J. Frishberg ARBITRARINESS AND ICONICITY: HISTORICAL CHANGE IN AMERICAN SIGN LANGUAGE , 1975 .

[30]  Mark Steedman,et al.  Animated conversation: rule-based generation of facial expression, gesture & spoken intonation for multiple conversational agents , 1994, SIGGRAPH.

[31]  J. Coates,et al.  Turn‐taking patterns in deaf conversation , 2001 .

[32]  Craig Martell FORM: An Extensible, Kinematically-based Gesture Annotation Scheme , 2002, LREC.

[33]  Ipke Wachsmuth,et al.  Gesture and Sign Language in Human-Computer Interaction , 1998, Lecture Notes in Computer Science.

[34]  R. Mitchell,et al.  How many deaf people are there in the United States? Estimates from the Survey of Income and Program Participation. , 2005, Journal of deaf studies and deaf education.

[35]  Richard Kennaway,et al.  Synthetic Animation of Deaf Signing Gestures , 2001, Gesture Workshop.

[36]  Gennaro Costagliola,et al.  Parsing Nonlinear Languages , 1994, IEEE Trans. Software Eng..

[37]  松本 崇,et al.  Analysis of News Sentences with Sign Language and Sign Translation Processing , 1992 .

[38]  H. Lane,et al.  A journey into the deaf-world , 1996 .

[39]  Éva Sáfár,et al.  The Architecture of an English-Text-to-Sign-Languages Translation System , 2001 .

[40]  Carol Neidle,et al.  The Syntax of American Sign Language: Functional Categories and Hierarchical Structure , 1999 .

[41]  Karen Emmorey,et al.  Perspectives on Classifier Constructions in Sign Languages , 2003 .

[42]  Matt Huenerfauth,et al.  A Multi-Path Architecture for Machine Translation of English Text into American Sign Language Animation , 2004, NAACL.

[43]  Jacob D. Furst,et al.  Database Design for American Sign Language , 2000 .

[44]  Alexander Franz,et al.  Large-scale machine translation: an interlingua approach , 1994, IEA/AIE '94.

[45]  Matt Huenerfauth Spatial Representation of Classifier Predicates for Machine Translation into American Sign Language , 2004 .

[46]  Ann Kathleen Farmer,et al.  On the interaction of morphology and syntax , 1980 .

[47]  George O. Strawn,et al.  Multidimensional Trees , 1991, Theor. Comput. Sci..

[48]  Carol Padden,et al.  American Sign Language and the architecture of phonological theory , 1987 .

[49]  Norman I. Badler,et al.  New Behavioral Paradigms for Virtual Human Models , 2005 .

[50]  Martha Palmer,et al.  Extending a Verb-lexicon Using a Semantically Annotated Corpus , 2004, LREC.

[51]  Margriet Verlinden,et al.  A Signing Avatar on the WWW , 2001, Gesture Workshop.

[52]  William Schuler,et al.  Using Model-Theoretic Semantic Interpretation to Guide Statistical Parsing and Word Recognition in a Spoken Language Interface , 2003, ACL.

[53]  Ted Suppalla,et al.  The Classifier System in American Sign Language , 1986 .

[54]  Catherine Pelachaud,et al.  Greta: A Simple Facial Animation Engine , 2002 .

[55]  Norman I. Badler,et al.  A machine translation system from English to American Sign Language , 2000, AMTA.

[56]  Matt Huenerfauth American Sign Language Generation: Multimodal NLG with Multiple Linguistic Channels , 2005, ACL.

[57]  Tomoko Sakiyama,et al.  Pattern recognition and synthesis for sign language translation system , 1994, Assets '94.

[58]  Linda Uyechi,et al.  Current Issues in ASL Phonology: Phonetics and Phonology , 1994 .

[59]  Manabu Okumura,et al.  Towards Automatic Translation from Japanese into Japanese Sign Language , 1998, Assistive Technology and Artificial Intelligence.

[60]  Matt Huenerfauth Representing coordination and non-coordination in an american sign language animation , 2005, Assets '05.

[61]  Norman I. Badler,et al.  Dynamically altering agent behaviors using natural language instructions , 2000, AGENTS '00.

[62]  Brian A. Barsky,et al.  Proceedings of the 12th Annual Conference on Computer Graphics and Interactive Techniques, SIGGRAPH 1985, San Francisco, California, USA, July 22-26, 1985 , 1985, SIGGRAPH.

[63]  Matt Huenerfauth,et al.  American Sign Language natural language generation and machine translation , 2005, ASAC.

[64]  Dean Barker,et al.  South African Sign Language Machine Translation System , 2003, AFRIGRAPH '03.

[65]  Rosalee Wolfe An interface for transcribing American Sign Language , 1999, SIGGRAPH '99.

[66]  Ying Liu,et al.  Applying Empirical Data on Upper Torso Movement to Real-time Collision-free Reach Tasks , 2005 .

[67]  Mark Liberman,et al.  A formal framework for linguistic annotation , 1999, Speech Commun..

[68]  Bonnie J. Dorr,et al.  The use of lexical semantics in interlingual machine translation , 2004, Machine Translation.

[69]  Justine Cassell,et al.  BEAT: the Behavior Expression Animation Toolkit , 2001, Life-like characters.

[70]  Kim Marriott,et al.  Towards a hierarchy of visual languages , 1996, Proceedings 1996 IEEE Symposium on Visual Languages.

[71]  Mark Wells,et al.  Virtual signing: capture, animation, storage and transmission-an overview of the ViSiCAST project , 2000 .

[72]  Shan Lu,et al.  Towards a Dialogue System Based on Recognition and Synthesis of Japanese Sign Language , 1997, Gesture Workshop.

[73]  Yue Cao,et al.  Total-Order Planning with Partially Ordered Subtasks , 2001, IJCAI.

[74]  Ken Shoemake,et al.  Animating rotation with quaternion curves , 1985, SIGGRAPH.

[75]  Brenda Schick,et al.  The acquisition of classifier predicates in American Sign Language , 1987 .

[76]  Wen Gao,et al.  Study on translating Chinese into Chinese sign language , 2008, Journal of Computer Science and Technology.

[77]  Matt Huenerfauth Representing coordination and non-coordination in American Sign Language animations , 2006, Behav. Inf. Technol..

[78]  Scott K. Liddell Grammar, Gesture, and Meaning in American Sign Language , 2003 .

[79]  Matt Huenerfauth,et al.  Spatial and planning models of ASL classifier predicates for machine translation , 2004, TMI.