Every Rating Matters: Joint Learning of Subjective Labels and Individual Annotators for Speech Emotion Classification
暂无分享,去创建一个
[1] Björn W. Schuller,et al. From Hard to Soft: Towards more Human-like Emotion Recognition by Modelling the Perception Uncertainty , 2017, ACM Multimedia.
[2] P A Turski,et al. Metabolic rate in the right amygdala predicts negative affect in depressed patients , 1998, Neuroreport.
[3] Yusuke Ijima,et al. Soft-Target Training with Ambiguous Emotional Utterances for DNN-Based Speech Emotion Classification , 2018, 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[4] Stefan Winkler,et al. ASCERTAIN: Emotion and Personality Recognition Using Commercial Sensors , 2018, IEEE Transactions on Affective Computing.
[5] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[6] Hsi-Pin Ma,et al. NNIME: The NTHU-NTUA Chinese interactive multimodal emotion corpus , 2017, 2017 Seventh International Conference on Affective Computing and Intelligent Interaction (ACII).
[7] S. Shott,et al. Emotion and Social Life: A Symbolic Interactionist Analysis , 1979, American Journal of Sociology.
[8] Margaret Lech,et al. Modeling subjectiveness in emotion recognition with deep neural networks: Ensembles vs soft labels , 2016, 2016 International Joint Conference on Neural Networks (IJCNN).
[9] Min Chen,et al. AIWAC: affective interaction through wearable computing and cloud technology , 2015, IEEE Wireless Communications.
[10] A. Bechara. The role of emotion in decision-making: Evidence from neurological patients with orbitofrontal damage , 2004, Brain and Cognition.
[11] Bir Bhanu,et al. To skip or not to skip? A dataset of spontaneous affective response of online advertising (SARA) for audience behavior analysis , 2015, 2015 11th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition (FG).
[12] Ming Yang,et al. DeepFace: Closing the Gap to Human-Level Performance in Face Verification , 2014, 2014 IEEE Conference on Computer Vision and Pattern Recognition.
[13] Carlos Busso,et al. IEMOCAP: interactive emotional dyadic motion capture database , 2008, Lang. Resour. Evaluation.
[14] Emily Mower Provost,et al. Leveraging inter-rater agreement for audio-visual emotion recognition , 2015, 2015 International Conference on Affective Computing and Intelligent Interaction (ACII).
[15] Sigal G. Barsade. The Ripple Effect: Emotional Contagion and its Influence on Group Behavior , 2002 .
[16] P. Wesley Schultz,et al. Values and Proenvironmental Behavior , 1998 .
[17] Andrew Zisserman,et al. Temporal HeartNet: Towards Human-Level Automatic Analysis of Fetal Cardiac Screening Video , 2017, MICCAI.
[18] Seyedmahdad Mirsamadi,et al. Automatic speech emotion recognition using recurrent neural networks with local attention , 2017, 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[19] Björn Schuller,et al. Opensmile: the munich versatile and fast open-source audio feature extractor , 2010, ACM Multimedia.
[20] Carlos Busso,et al. Interpreting ambiguous emotional expressions , 2009, 2009 3rd International Conference on Affective Computing and Intelligent Interaction and Workshops.
[21] Geoffrey E. Hinton,et al. Who Said What: Modeling Individual Labelers Improves Classification , 2017, AAAI.