Enhancing Mouth-Based Emotion Recognition Using Transfer Learning

This work concludes the first study on mouth-based emotion recognition while adopting a transfer learning approach. Transfer learning results are paramount for mouth-based emotion emotion recognition, because few datasets are available, and most of them include emotional expressions simulated by actors, instead of adopting real-world categorisation. Using transfer learning, we can use fewer training data than training a whole network from scratch, and thus more efficiently fine-tune the network with emotional data and improve the convolutional neural network’s performance accuracy in the desired domain. The proposed approach aims at improving emotion recognition dynamically, taking into account not only new scenarios but also modified situations to the initial training phase, because the image of the mouth can be available even when the whole face is visible only in an unfavourable perspective. Typical applications include automated supervision of bedridden critical patients in a healthcare management environment, and portable applications supporting disabled users having difficulties in seeing or recognising facial emotions. This achievement takes advantage of previous preliminary works on mouth-based emotion recognition using deep-learning, and has the further benefit of having been tested and compared to a set of other networks using an extensive dataset for face-based emotion recognition, well known in the literature. The accuracy of mouth-based emotion recognition was also compared to the corresponding full-face emotion recognition; we found that the loss in accuracy is mostly compensated by consistent performance in the visual emotion recognition domain. We can, therefore, state that our method proves the importance of mouth detection in the complex process of emotion recognition.

[1]  Osvaldo Gervasi,et al.  A Method for Predicting Words by Interpreting Labial Movements , 2016, ICCSA.

[2]  Stefanos Zafeiriou,et al.  A Semi-automatic Methodology for Facial Landmark Annotation , 2013, 2013 IEEE Conference on Computer Vision and Pattern Recognition Workshops.

[3]  Daniele Nardi,et al.  Speaky for robots: the development of vocal interfaces for robotic applications , 2015, Applied Intelligence.

[4]  PanticMaja,et al.  300 Faces In-The-Wild Challenge , 2016 .

[5]  Sergey Ioffe,et al.  Rethinking the Inception Architecture for Computer Vision , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[6]  Jordi Vallverdú,et al.  Emotional machines: The next revolution , 2019, Web Intell..

[7]  Xiaowei Gu,et al.  Cybernetics of the Mind: Learning Individual's Perceptions Autonomously , 2017, IEEE Systems, Man, and Cybernetics Magazine.

[8]  Geoffrey E. Hinton,et al.  ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.

[9]  Josephine Sullivan,et al.  One millisecond face alignment with an ensemble of regression trees , 2014, 2014 IEEE Conference on Computer Vision and Pattern Recognition.

[10]  Osvaldo Gervasi,et al.  Towards a Learning-Based Performance Modeling for Accelerating Deep Neural Networks , 2019, ICCSA.

[11]  Sergey Ioffe,et al.  Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning , 2016, AAAI.

[12]  Magy Seif El-Nasr,et al.  Eliciting Emotions in Design of Games - a Theory Driven Approach , 2016, EMPIRE@RecSys.

[13]  Jennifer Healey,et al.  Toward Machine Emotional Intelligence: Analysis of Affective Physiological State , 2001, IEEE Trans. Pattern Anal. Mach. Intell..

[14]  Jordi Vallverdú,et al.  Emotional affordances in human-machine interactive planning and negotiation , 2017, WI.

[15]  Andrew Zisserman,et al.  Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.

[16]  Osvaldo Gervasi,et al.  EmEx, a Tool for Automated Emotive Face Recognition Using Convolutional Neural Networks , 2017, ICCSA.

[17]  Stefanos Zafeiriou,et al.  300 Faces in-the-Wild Challenge: The First Facial Landmark Localization Challenge , 2013, 2013 IEEE International Conference on Computer Vision Workshops.

[18]  M. Pantic,et al.  Faces InThe-Wild Challenge : Database and Results , 2016 .

[19]  Mohammad H. Mahoor,et al.  AffectNet: A Database for Facial Expression, Valence, and Arousal Computing in the Wild , 2017, IEEE Transactions on Affective Computing.

[20]  Valentina Franzoni,et al.  Web-Based Similarity for Emotion Recognition in Web Objects , 2016, 2016 IEEE/ACM 9th International Conference on Utility and Cloud Computing (UCC).

[21]  Camille Couprie,et al.  Learning Hierarchical Features for Scene Labeling , 2013, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[22]  François Chollet,et al.  Xception: Deep Learning with Depthwise Separable Convolutions , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[23]  Fatih Uzdilli,et al.  Potential and Limitations of Commercial Sentiment Detection Tools , 2013, ESSEM@AI*IA.

[24]  Guigang Zhang,et al.  Deep Learning , 2016, Int. J. Semantic Comput..

[25]  Viviana Patti,et al.  Emotional Responses to Artworks in Online Collections , 2013, UMAP Workshops.

[26]  Valentina Franzoni,et al.  Web-based Semantic Similarity for Emotion Recognition in Web Objects , 2016, ArXiv.

[27]  Valentina Franzoni,et al.  Errors, Biases and Overconfidence in Artificial Emotional Modeling , 2019, WI.

[28]  Li Fei-Fei,et al.  ImageNet: A large-scale hierarchical image database , 2009, CVPR.

[29]  Robert Plutchik,et al.  A psychoevolutionary theory of emotions , 1982 .

[30]  Osvaldo Gervasi,et al.  Automating facial emotion recognition , 2019, Web Intell..

[31]  Rosalind W. Picard Affective computing: challenges , 2003, Int. J. Hum. Comput. Stud..

[32]  Osvaldo Gervasi,et al.  An Approach for Improving Automatic Mouth Emotion Recognition , 2019, ICCSA.

[33]  Osvaldo Gervasi,et al.  A Brain Computer Interface for Enhancing the Communication of People with Severe Impairment , 2014, ICCSA.

[34]  Amos J. Storkey,et al.  Data Augmentation Generative Adversarial Networks , 2017, ICLR 2018.

[35]  Stefanos Zafeiriou,et al.  300 Faces In-The-Wild Challenge: database and results , 2016, Image Vis. Comput..

[36]  P. Ekman An argument for basic emotions , 1992 .

[37]  Valentina Franzoni,et al.  Emotion Recognition for Self-aid in Addiction Treatment, Psychotherapy, and Nonviolent Communication , 2019, ICCSA.

[38]  Gillian R. Hayes,et al.  Interactive visual supports for children with autism , 2010, Personal and Ubiquitous Computing.