Research and Analysis of Learning Rate on Face Expression Recognition

As an important hyperparameter in facial expression recognition, learning rate determines whether the facial expression recognition function can converge to the local minimum and when it converges to the minimum value. With the increasing complexity of the facial expression recognition network model, it is particularly difficult to find a suitable learning rate, because the size of different parameters varies greatly and needs to be adjusted throughout the training process, which not only consumes a lot of time. Sometimes, it is not possible to adjust to the optimal, which makes it difficult to improve the accuracy of facial expression recognition. Therefore, in this paper, through the research and analysis of the facial expression recognition rate, we propose face expression recognition based on the Adabound algorithm and compare it with SGD (Stochastic Gradient Descent) and Adam on the face expression recognition network model use CK+ (Cohn-Kanade) dataset. The Adabound adaptive optimization algorithm achieved high accuracy of 96.97% and 100% on ResNet18 and ResNet34 respectively. Compared with recent articles, our model based on the Adabound algorithm has reached a very high level in the CK+ dataset in terms of the effect of a single model.

[1]  Sanjiv Kumar,et al.  On the Convergence of Adam and Beyond , 2018 .

[2]  Zhiyong Feng,et al.  Facial expression recognition via deep learning , 2014, 2014 International Conference on Smart Computing.

[3]  Thomas S. Huang,et al.  Do Deep Neural Networks Learn Facial Action Units When Doing Expression Recognition? , 2015, 2015 IEEE International Conference on Computer Vision Workshop (ICCVW).

[4]  Yoram Singer,et al.  Adaptive Subgradient Methods for Online Learning and Stochastic Optimization , 2011, J. Mach. Learn. Res..

[5]  Jian Sun,et al.  Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[6]  Ping Liu,et al.  Identity-Aware Convolutional Neural Network for Facial Expression Recognition , 2017, 2017 12th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2017).

[7]  Matthew D. Zeiler ADADELTA: An Adaptive Learning Rate Method , 2012, ArXiv.

[8]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[9]  Ping Liu,et al.  Facial Expression Recognition via a Boosted Deep Belief Network , 2014, 2014 IEEE Conference on Computer Vision and Pattern Recognition.

[10]  Nathan Srebro,et al.  The Marginal Value of Adaptive Gradient Methods in Machine Learning , 2017, NIPS.

[11]  Xu Sun,et al.  Adaptive Gradient Methods with Dynamic Bound of Learning Rate , 2019, ICLR.