In this paper, we present an approach for building translation memory for American Sign Language (ASL) from parallel corpora between English and ASL, by identifying new alignment combinations of words from existing texts. Our primary contribution is the application of several models of alignments for Sign Language. The model represents probabilistic relationships between properties of words, and relates them to learned underlying causes of structural variability within the domain. We developed a statistical machine translation based on generated translation memory. The model was evaluated on a big parallel corpus containing more than 800 millions of words. IBM Models have been applied to align Sign Language Corpora then we have run experimentation on a big collection of paired data between English and American Sign Language. The result is useful to build a Statistical Machine Language or any related field.
[1]
Andy Way,et al.
Joining Hands: Developing A Sign Language Machine Translation System with and for the Deaf Community
,
2007,
CVHI.
[2]
Robert L. Mercer,et al.
The Mathematics of Statistical Machine Translation: Parameter Estimation
,
1993,
CL.
[3]
Mohamed Jemni,et al.
A Web-Based Tool to Create Online Courses for Deaf Pupils
,
2007
.
[4]
Mohamed Jemni,et al.
English-ASL Gloss Parallel Corpus 2012: ASLG-PC12
,
2012
.
[5]
Daniel Marcu,et al.
Statistical Phrase-Based Translation
,
2003,
NAACL.
[6]
Philip Koehn,et al.
Statistical Machine Translation
,
2010,
EAMT.
[7]
Ian W. Marshall,et al.
Sign language generation using HPSG.
,
2002,
TMI.
[8]
Matt Huenerfauth,et al.
Evaluation of American Sign Language Generation by Native ASL Signers
,
2008,
TACC.