Grammatical facial expression recognition using customized deep neural network architecture

This paper proposes to expand the visual understanding capacity of computers by helping it recognize human sign language more efficiently. This is carried out through recognition of facial expressions, which accompany the hand signs used in this language. This paper specially focuses on the popular Brazilian sign language (LIBRAS). While classifying different hand signs into their respective word meanings has already seen much literature dedicated to it, the emotions or intention with which the words are expressed haven't primarily been taken into consideration. As from our normal human experience, words expressed with different emotions or mood can have completely different meanings attached to it. Lending computers the ability of classifying these facial expressions, can help add another level of deep understanding of what the deaf person exactly wants to communicate. The proposed idea is implemented through a deep neural network having a customized architecture. This helps learning specific patterns in individual expressions much better as compared to a generic approach. With an overall accuracy of 98.04%, the implemented deep network performs excellently well and thus is fit to be used in any given practical scenario.

[1]  Benjamín R. C. Bedregal,et al.  Fuzzy Rule-Based Hand Gesture Recognition , 2006, IFIP AI.

[2]  Yoshua Bengio,et al.  Understanding the difficulty of training deep feedforward neural networks , 2010, AISTATS.

[3]  Angelo C. Loula,et al.  Recognition of Static Gestures Applied to Brazilian Sign Language (Libras) , 2015, 2015 28th SIBGRAPI Conference on Graphics, Patterns and Images.

[4]  Ednaldo Brigante Pizzolato,et al.  Sign Language Recognition with Support Vector Machines and Hidden Conditional Random Fields: Going from Fingerspelling to Natural Articulated Words , 2013, MLDM.

[5]  Daniel Kelly,et al.  A framework for continuous multimodal sign language recognition , 2009, ICMI-MLMI '09.

[6]  Helton Hideraldo Bíscaro,et al.  Hand movement recognition for Brazilian Sign Language: A study using distance-based neural networks , 2009, 2009 International Joint Conference on Neural Networks.

[7]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[8]  Hemerson Pistori,et al.  An Experiment on Handshape Sign Recognition Using Adaptive Technology: Preliminary Results , 2004, SBIA.

[9]  Ismail Ari,et al.  Facial feature tracking and expression recognition for sign language , 2009, 2009 IEEE 17th Signal Processing and Communications Applications Conference.

[10]  Moritz Knorr,et al.  The significance of facial features for automatic sign language recognition , 2008, 2008 8th IEEE International Conference on Automatic Face & Gesture Recognition.

[11]  Sebastian Feuerstack,et al.  A real-time system to recognize static gestures of Brazilian sign language (libras) alphabet using Kinect , 2012, IHC.

[12]  Mauro dos Santos Anjo,et al.  Automatic recognition of finger spelling for LIBRAS based on a two-layer architecture , 2010, SAC '10.

[13]  Zdenek Krnoul,et al.  Correlation analysis of facial features and sign gestures , 2010, IEEE 10th INTERNATIONAL CONFERENCE ON SIGNAL PROCESSING PROCEEDINGS.

[14]  Luiz Eduardo Soares de Oliveira,et al.  LIBRAS Sign Language Hand Configuration Recognition Based on 3D Meshes , 2013, 2013 IEEE International Conference on Systems, Man, and Cybernetics.

[15]  Sarajane Marques Peres,et al.  Grammatical Facial Expressions Recognition with Machine Learning , 2014, FLAIRS Conference.