This paper provides a survey literature survey on the emotion recognition in spoken dialogs and proposes an implementation of such a system using acoustic features. The data corpus contains 322 utterances expressing four emotions such as happy, angry, sad, and fear. 50% of the total data is used for training while the other 50% is used for testing. We use 21 features extracted from our features set in our experiment. The feature vectors are normalized by using Z-score normalization. The multi-class support vector machine (SVM) classifier is used for classification. The result shows that sad is classified with the highest accuracy whereas happy is classified with the least accuracy
[1]
Oh-Wook Kwon,et al.
EMOTION RECOGNITION BY SPEECH SIGNAL
,
2003
.
[2]
Jiucang Hao,et al.
Emotion recognition by speech signals
,
2003,
INTERSPEECH.
[3]
Shrikanth S. Narayanan,et al.
Toward detecting emotions in spoken dialogs
,
2005,
IEEE Transactions on Speech and Audio Processing.
[4]
Eric Keller,et al.
Prosodic aspects of speech
,
1995
.
[5]
Janet M Beck,et al.
Hearing smiles - Perceptual, acoustic and production aspects of labial spreading.
,
1999
.
[6]
Valery A. Petrushin,et al.
EMOTION IN SPEECH: RECOGNITION AND APPLICATION TO CALL CENTERS
,
1999
.