Emotions play an important role in people's life. Understanding and recognising is not only important for interpersonal communication, but also has promising applications in Human-Computer Interaction, automobile safety and medical research. This project focuses on extending the emotion recognition database, and training the CNN + RNN emotion recognition neural networks with emotion category representation and valence \& arousal representation. The combined models are constructed by training the two representations simultaneously. The comparison and analysis between the three types of model are discussed. The inner-relationship between two emotion representations and the interpretability of the neural networks are investigated. The findings suggest that categorical emotion recognition performance can benefit from training with a combined model. And the mapping of emotion category and valence \& arousal values can explain this phenomenon.
[1]
Geoffrey E. Hinton,et al.
Visualizing Data using t-SNE
,
2008
.
[2]
Guoying Zhao,et al.
Aff-Wild: Valence and Arousal ‘In-the-Wild’ Challenge
,
2017,
2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).
[3]
Jubilant J. Kizhakkethottam,et al.
Overview on emotion recognition system
,
2015,
2015 International Conference on Soft-Computing and Networks Security (ICSNS).
[4]
Kai Wang,et al.
Group emotion recognition with individual facial emotion CNNs and global image based CNNs
,
2017,
ICMI.
[5]
Maja Pantic,et al.
Fully Automatic Recognition of the Temporal Phases of Facial Actions
,
2012,
IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics).
[6]
Yi-Hsuan Yang,et al.
Machine Recognition of Music Emotion: A Review
,
2012,
TIST.