Improving and Extending Continuous Sign Language Recognition: Taking Iconicity and Spatial Language into account

In a lot of recent research, attention has been drawn to recognizing sequences of lexical signs in continuous Sign Language corpora, often artificial. However, as SLs are structured through the use of space and iconicity, focusing on lexicon only prevents the field of Continuous Sign Language Recognition (CSLR) from extending to Sign Language Understanding and Translation. In this article, we propose a new formulation of the CSLR problem and discuss the possibility of recognizing higher-level linguistic structures in SL videos, like classifier constructions. These structures show much more variability than lexical signs, and are fundamentally different than them in the sense that form and meaning can not be disentangled. Building on the recently published French Sign Language corpus Dicta-Sign-LSF-v2, we also discuss the performance and relevance of a simple recurrent neural network trained to recognize illustrative structures.

[1]  Changshui Zhang,et al.  A Deep Neural Framework for Continuous Sign Language Recognition by Iterative Training , 2019, IEEE Transactions on Multimedia.

[2]  Charlotte Lee Baker-Shenk,et al.  American Sign Language : A Teacher's Resource Text on Grammar and Culture , 1991 .

[3]  Sylvain Kahane,et al.  L’anatomie du proforme en langue des signes française : , 2018, TIPA. Travaux interdisciplinaires sur la parole et le langage.

[4]  Michèle Gouiffès,et al.  Dicta-Sign-LSF-v2: Remake of a Continuous French Sign Language Dialogue Corpus and a First Baseline for Automatic Sign Language Processing , 2020, LREC.

[5]  E. F. K. Ko AMSTERDAM STUDIES IN THE THEORY AND HISTORY OF LINGUISTIC SCIENCE , 2006 .

[6]  Christian Cuxac,et al.  La langue des signes francaise (LSF) : les voies de l'iconicité , 2000 .

[7]  Trevor Johnston,et al.  Auslan Corpus Annotation Guidelines , 2013 .

[8]  Karl-Friedrich Kraiss,et al.  Towards a Video Corpus for Signer-Independent Continuous Sign Language Recognition , 2007 .

[9]  Jie Huang,et al.  Video-based Sign Language Recognition without Temporal Segmentation , 2018, AAAI.

[10]  Hermann Ney,et al.  Extensions of the Sign Language Recognition and Translation Corpus RWTH-PHOENIX-Weather , 2014, LREC.

[11]  Hermann Ney,et al.  Re-Sign: Re-Aligned End-to-End Sequence Modelling with Deep Recurrent CNN-HMMs , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[12]  Adam Schembri,et al.  Australian Sign Language: Auslan: An Introduction to Sign Language Linguistics , 2007 .

[13]  Houqiang Li,et al.  Iterative Alignment Network for Continuous Sign Language Recognition , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[14]  Patricia Siple,et al.  Understanding language through sign language research , 1978 .

[15]  John Glauert,et al.  Dicta-Sign – Building a Multilingual Sign Language Corpus , 2012 .