SILFA: Sign Language Facial Action Database for the Development of Assistive Technologies for the Deaf

Facial expressions are fundamental in Sign Languages (SLs), which are visuospatial linguistic systems structured on gestures, adopted around the world by deaf people to communicate. Deaf individuals frequently need a sign language interpreter in their access to school and public services. In such scenarios, the absence of interpreters typically results in discouraging experiences. Developments in Automatic Sign Language Recognition (ASLR) can enable newer assistive technologies and change the interaction of the deaf with the world. One major barrier to improving ASLR is the difficulty in obtaining sets of well-annotated data. We present a newly developed video database of Brazilian Sign Language facial expressions in a diverse group of deaf and hearing young adults. Well-validated sentences stimulus were used to elicit affective and grammatical facial expressions. Frame-work ground-truth for facial actions was manually annotated using the Facial Action Coding System (FACS). Also, the work promotes the exploration of discriminant features in subtle facial expression in sign language, a better understanding of the relation between grammatical facial expression class dynamics in facial action units, and a deeper understanding of its facial action occurrence. To provide a baseline for use in future research, protocols and benchmarks for automated action unit recognition are reported.

[1]  P. Ekman,et al.  Measuring facial movement , 1976 .

[2]  Patrice Y. Simard,et al.  Best practices for convolutional neural networks applied to visual document analysis , 2003, Seventh International Conference on Document Analysis and Recognition, 2003. Proceedings..

[3]  King-Sun Fu,et al.  IEEE Transactions on Pattern Analysis and Machine Intelligence Publication Information , 2004, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[4]  Surendra Ranganath,et al.  Automatic Sign Language Analysis: A Survey and the Future beyond Lexical Meaning , 2005, IEEE Trans. Pattern Anal. Mach. Intell..

[5]  Andrea L. Berez Review of EUDICO Linguistic Annotator (ELAN) , 2007 .

[6]  Moritz Knorr,et al.  The significance of facial features for automatic sign language recognition , 2008, 2008 8th IEEE International Conference on Automatic Face & Gesture Recognition.

[7]  Davis E. King,et al.  Dlib-ml: A Machine Learning Toolkit , 2009, J. Mach. Learn. Res..

[8]  Maja Pantic,et al.  A Dynamic Texture-Based Approach to Recognition of Facial Actions and Their Temporal Models , 2010, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[9]  Richard Bowden,et al.  Sign Language Recognition , 2011, Visual Analysis of Humans.

[10]  Thad Starner,et al.  American sign language recognition with the kinect , 2011, ICMI '11.

[11]  R. Gur,et al.  Automated Facial Action Coding System for dynamic analysis of facial expressions in neuropsychiatric disorders , 2011, Journal of Neuroscience Methods.

[12]  Geoffrey E. Hinton,et al.  ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.

[13]  Ronnie B. Wilbur,et al.  Discriminant Features and Temporal Structure of Nonmanuals in American Sign Language , 2014, PloS one.

[14]  Sarajane Marques Peres,et al.  Grammatical Facial Expressions Recognition with Machine Learning , 2014, FLAIRS Conference.

[15]  Fei Yang,et al.  Non-manual grammatical marker recognition based on multi-scale, spatio-temporal analysis of head pose and facial expressions , 2014, Image Vis. Comput..

[16]  Benjamin Schrauwen,et al.  Sign Language Recognition Using Convolutional Neural Networks , 2014, ECCV Workshops.

[17]  Hermann Ney,et al.  Continuous sign language recognition: Towards large vocabulary statistical recognition systems handling multiple signers , 2015, Comput. Vis. Image Underst..

[18]  Andrew Zisserman,et al.  Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.

[19]  Martin Kampel,et al.  Facial Expression Recognition using Convolutional Neural Networks: State of the Art , 2016, ArXiv.

[20]  José Mario De Martino,et al.  Brazilian Sign Language Recognition Using Kinect , 2016, ECCV Workshops.

[21]  Lale Akarun,et al.  BosphorusSign: A Turkish Sign Language Recognition Corpus in Health and Finance Domains , 2016, LREC.

[22]  Fernando De la Torre,et al.  Learning Spatial and Temporal Cues for Multi-Label Facial Action Unit Detection , 2017, 2017 12th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2017).

[23]  Emely Pujólli da Silva,et al.  Recognition of Non-Manual Expressions in Brazilian Sign Language , 2017 .

[24]  Wen-Sheng Chu,et al.  Learning Facial Action Units from Web Images with Scalable Weakly Supervised Clustering , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[25]  Masahiko Toyonaga,et al.  Facial Expression Sequence Recognition for a Japanese Sign Language Training System , 2018, 2018 Joint 10th International Conference on Soft Computing and Intelligent Systems (SCIS) and 19th International Symposium on Advanced Intelligent Systems (ISIS).

[26]  Lijun Yin,et al.  EAC-Net: Deep Nets with Enhancing and Cropping for Facial Action Unit Detection , 2018, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[27]  Maja Pantic,et al.  Automatic Analysis of Facial Actions: A Survey , 2019, IEEE Transactions on Affective Computing.

[28]  Publishing DGS corpus data: Different Formats for Different Needs , .