Towards Explaining Deep Learning Networks to Distinguish Facial Expressions of Pain and Emotions

Deep learning networks are successfully used for object and face recognition in images and videos. In order to be able to apply such networks in practice, for example in hospitals as a pain recognition tool, the current procedures are only suitable to a limited extent. The advantage of deep learning methods is that they can learn complex non-linear relationships between raw data and target classes without limiting themselves to a set of hand-crafted features provided by humans. However, the disadvantage is that due to the complexity of these networks, it is not possible to interpret the knowledge that is stored inside the network. It is a black-box learning procedure. Explainable Artificial Intelligence (AI) approaches mitigate this problem by extracting explanations for decisions and representing them in a human-interpretable form. The aim of this paper is to investigate the explainable AI method Layer-wise Relevance Propagation (LRP) and apply it to explain how a deep learning network distinguishes facial expressions of pain from facial expressions of emotions such as happiness and disgust.

[1]  D. Hubel,et al.  Receptive fields, binocular interaction and functional architecture in the cat's visual cortex , 1962, The Journal of physiology.

[2]  Jürgen Schmidhuber,et al.  Deep learning in neural networks: An overview , 2014, Neural Networks.

[3]  Y. Trope,et al.  Body Cues, Not Facial Expressions, Discriminate Between Intense Positive and Negative Emotions , 2012, Science.

[4]  Ayoub Al-Hamadi,et al.  The biovid heat pain database data for the advancement and systematic validation of an automated pain recognition system , 2013, 2013 IEEE International Conference on Cybernetics (CYBCO).

[5]  Guigang Zhang,et al.  Deep Learning , 2016, Int. J. Semantic Comput..

[6]  N. Ambady,et al.  Thin slices of expressive behavior as predictors of interpersonal consequences: A meta-analysis. , 1992 .

[7]  Alexander Binder,et al.  Controlling explanatory heatmap resolution and semantics via decomposition depth , 2016, 2016 IEEE International Conference on Image Processing (ICIP).

[8]  Chirag Ravat,et al.  Survey on Different Methods to Improve Accuracy of The Facial Expression Recognition Using Artificial Neural Networks , 2018 .

[9]  C. Frith Role of facial expressions in social interactions , 2009, Philosophical Transactions of the Royal Society B: Biological Sciences.

[10]  Jeffrey F. Cohn,et al.  Painful data: The UNBC-McMaster shoulder pain expression archive database , 2011, Face and Gesture 2011.

[11]  Andrew Zisserman,et al.  Deep Face Recognition , 2015, BMVC.

[12]  Alexander Binder,et al.  Understanding and Comparing Deep Neural Networks for Age and Gender Classification , 2017, 2017 IEEE International Conference on Computer Vision Workshops (ICCVW).

[13]  P. Ekman,et al.  What the face reveals : basic and applied studies of spontaneous expression using the facial action coding system (FACS) , 2005 .

[14]  Wojciech Samek,et al.  Methods for interpreting and understanding deep neural networks , 2017, Digit. Signal Process..

[15]  Alexander Binder,et al.  On Pixel-Wise Explanations for Non-Linear Classifier Decisions by Layer-Wise Relevance Propagation , 2015, PloS one.

[16]  Ayoub Al-Hamadi,et al.  Automatic Pain Assessment with Facial Activity Descriptors , 2017, IEEE Transactions on Affective Computing.

[17]  Sheryl Brahnam,et al.  Machine recognition and representation of neonatal facial displays of acute pain , 2006, Artif. Intell. Medicine.