Analysis and Compensation of the Reaction Lag of Evaluators in Continuous Emotional Annotations
暂无分享,去创建一个
[1] Maja Pantic,et al. This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. IEEE TRANSACTIONS ON AFFECTIVE COMPUTING , 2022 .
[2] Björn W. Schuller,et al. Data-driven clustering in emotional space for affect recognition using discriminatively trained LSTM networks , 2009, INTERSPEECH.
[3] Roddy Cowie,et al. Real life emotions in French and English TV video clips: an integrated annotation protocol combining continuous and discrete approaches , 2006, LREC.
[4] Carlos Busso,et al. Energy and F0 contour modeling with functional data analysis for emotional speech detection , 2013, INTERSPEECH.
[5] Björn Schuller,et al. Opensmile: the munich versatile and fast open-source audio feature extractor , 2010, ACM Multimedia.
[6] Eckart Altenmüller,et al. EMuJoy: Software for continuous measurement of perceived emotions in music , 2007, Behavior research methods.
[7] Roddy Cowie,et al. Describing the emotional states that are expressed in speech , 2003, Speech Commun..
[8] Mohamed Chetouani,et al. Robust continuous prediction of human emotions using multiscale dynamic cues , 2012, ICMI '12.
[9] Nadia Bianchi-Berthouze,et al. Naturalistic Affective Expression Classification by a Multi-stage Approach Based on Hidden Markov Models , 2011, ACII.
[10] P. Ekman,et al. Facial action coding system: a technique for the measurement of facial movement , 1978 .
[11] Zhu Liang Yu,et al. Speech Emotion Recognition System Based on L1 Regularized Linear Regression and Decision Fusion , 2011, ACII.
[12] Carlos Busso,et al. Unveiling the Acoustic Properties that Describe the Valence Dimension , 2012, INTERSPEECH.
[13] Roddy Cowie,et al. Tracing Emotion: An Overview , 2012, Int. J. Synth. Emot..
[14] Klaus R. Scherer,et al. Vocal communication of emotion: A review of research paradigms , 2003, Speech Commun..
[15] Carlos Busso,et al. Modeling mutual influence of interlocutor emotion states in dyadic spoken interactions , 2009, INTERSPEECH.
[16] Youngmoo E. Kim,et al. MoodSwings: A Collaborative Game for Music Mood Label Collection , 2008, ISMIR.
[17] Björn W. Schuller,et al. Context-Sensitive Learning for Enhanced Audiovisual Emotion Classification , 2012, IEEE Transactions on Affective Computing.
[18] E. Nöth,et al. Releasing a thoroughly annotated and processed spontaneous emotional database : the FAU Aibo Emotion Corpus , 2008 .
[19] Björn W. Schuller,et al. The INTERSPEECH 2011 Speaker State Challenge , 2011, INTERSPEECH.
[20] Vladimir Pavlovic,et al. Dynamic Probabilistic CCA for Analysis of Affective Behaviour , 2012, ECCV.
[21] Gwen Littlewort,et al. Automatic Recognition of Facial Actions in Spontaneous Expressions , 2006, J. Multim..
[22] Dirk Heylen,et al. The Sensitive Artificial Listner: an induction technique for generating emotionally coloured conversation , 2008 .
[23] Roddy Cowie,et al. FEELTRACE: an instrument for recording perceived emotion in real time , 2000 .
[24] Ian H. Witten,et al. The WEKA data mining software: an update , 2009, SKDD.
[25] Angeliki Metallinou,et al. Annotation and processing of continuous emotional attributes: Challenges and opportunities , 2013, 2013 10th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition (FG).
[26] Mark A. Clements,et al. Investigating the Use of Formant Based Features for Detection of Affective Dimensions in Speech , 2011, ACII.
[27] Shrikanth Narayanan,et al. Toward Effective Automatic Recognition Systems of Emotion in Speech , 2014 .
[28] Hatice Gunes,et al. Continuous Prediction of Spontaneous Affect from Multiple Cues and Modalities in Valence-Arousal Space , 2011, IEEE Transactions on Affective Computing.
[29] Athanasia Zlatintsi,et al. A supervised approach to movie emotion tracking , 2011, 2011 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).