Verifying Deep Learning-based Decisions for Facial Expression Recognition

Neural networks with high performance can still be biased towards non-relevant features. However, reliability and robustness is especially important for high-risk fields such as clinical pain treatment. We therefore propose a verification pipeline, which consists of three steps. First, we classify facial expressions with a neural network. Next, we apply layer-wise relevance propagation to create pixel-based explanations. Finally, we quantify these visual explanations based on a bounding-box method with respect to facial regions. Although our results show that the neural network achieves state-of-the-art results, the evaluation of the visual explanations reveals that relevant facial regions may not be considered.

[1]  Jeffrey F. Cohn,et al.  Painful data: The UNBC-McMaster shoulder pain expression archive database , 2011, Face and Gesture 2011.

[2]  Alexander Binder,et al.  On Pixel-Wise Explanations for Non-Linear Classifier Decisions by Layer-Wise Relevance Propagation , 2015, PloS one.

[3]  Shinichi Nakajima,et al.  Towards Best Practice in Explaining Neural Network Decisions with LRP , 2019, 2020 International Joint Conference on Neural Networks (IJCNN).

[4]  Christian Küblbeck,et al.  Face detection and tracking in video sequences using the modifiedcensus transformation , 2006, Image Vis. Comput..

[5]  Takeo Kanade,et al.  The Extended Cohn-Kanade Dataset (CK+): A complete dataset for action unit and emotion-specified expression , 2010, 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition - Workshops.

[6]  Klaus-Robert Müller,et al.  Resolving challenges in deep learning-based analyses of histopathological images using explanation methods , 2019, Scientific Reports.

[7]  Klaus-Robert Müller,et al.  iNNvestigate neural networks! , 2018, J. Mach. Learn. Res..

[8]  Jian Sun,et al.  Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[9]  Shangfei Wang,et al.  Weakly Supervised Facial Action Unit Recognition Through Adversarial Training , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[10]  Jane Yung-jen Hsu,et al.  IdenNet: Identity-Aware Facial Action Unit Detection , 2019, 2019 14th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2019).

[11]  Peter Robinson,et al.  OpenFace: An open source facial behavior analysis toolkit , 2016, 2016 IEEE Winter Conference on Applications of Computer Vision (WACV).

[12]  Marcello Mortillaro,et al.  Emotion Expression from Different Angles: A Video Database for Facial Expressions of Actors Shot by a Camera Array , 2019, 2019 8th International Conference on Affective Computing and Intelligent Interaction (ACII).

[13]  Miriam Kunz,et al.  Facial muscle movements encoding pain—a systematic review , 2018, Pain.

[14]  Thomas Hauenstein,et al.  Towards Real-Time Head Pose Estimation: Exploring Parameter-Reduced Residual Networks on In-the-wild Datasets , 2019, IEA/AIE.