Numerous examinations are performed related to automatic emotion recognition and speech detection in the Laboratory of Speech Acoustics. This article reviews results achieved for automatic emotion recognition experiments on spontaneous speech databases on the base of the acoustical information only. Different acoustic parameters were compared for the acoustical preprocessing, and Support Vector Machines were used for the classification. In spontaneous speech, before the automatic emotion recognition, speech detection and speech segmentation are needed to segment the audio material into the unit of recognition. At present, phrase was selected as a unit of segmentation. A special method was developed on the base of Hidden Markov Models, which can process the speech detection and automatic phrase segmentation simultaneously. The developed method was tested in a noisy spontaneous telephone speech database. The emotional classification was prepared on the detected and segmented speech.
[1]
Klára Vicsi,et al.
Speech Emotion Perception by Human and Machine
,
2008,
COST 2102 Workshop.
[2]
Chih-Jen Lin,et al.
LIBSVM: A library for support vector machines
,
2011,
TIST.
[3]
Inma Hernáez,et al.
An objective and subjective study of the role of semantics and prosodic features in building corpora for emotional TTS
,
2006,
IEEE Transactions on Audio, Speech, and Language Processing.
[4]
Paul Boersma,et al.
Praat, a system for doing phonetics by computer
,
2002
.
[5]
Zdravko Kacic,et al.
A rule-based emotion-dependent feature extraction method for emotion analysis from speech.
,
2006,
The Journal of the Acoustical Society of America.