Relative amplitude based features for emotion detection from speech

Emotion detection from speech has been realized to provide benefits for more natural human-machine interaction. To detect the emotion from speech signal, an abundantly long continuous speech segment is needed. This paper proposed a navel approach for emotion detection based on relative amplitude of speech signal. Relative amplitude reduces bias of glottis mutation of speech wave amplitude and obtains a normalized measure without concern of information from being distinct in feature. RBFC approach is used for segmentation of speech signal and the results are compared with other voiced segmentation approaches. Berlin emotional speech database is used for experimental purpose. The results show that accurate emotion recognition is obtained with optimum length of emotional speech. The RBFC features generate more accurate results than the other methods.

[1]  Mohamed Rizon Discrete Wavelet Transform Based Classification of Human Emotions Using Electroencephalogram Signals , 2010 .

[2]  Jiucang Hao,et al.  Emotion recognition by speech signals , 2003, INTERSPEECH.

[3]  E. Leon,et al.  Towards a robust real-time emotion detection system for intelligent buildings , 2005 .

[4]  Ling Guan,et al.  A neural network approach for human emotion recognition in speech , 2004, 2004 IEEE International Symposium on Circuits and Systems (IEEE Cat. No.04CH37512).

[5]  Yi Luo,et al.  Speech emotion recognition based on a hybrid of HMM/ANN , 2007 .

[6]  G. Castellanos-Dominguez,et al.  Feature extraction of speech signals in emotion identification , 2008, 2008 30th Annual International Conference of the IEEE Engineering in Medicine and Biology Society.

[7]  Oh-Wook Kwon,et al.  EMOTION RECOGNITION BY SPEECH SIGNAL , 2003 .

[8]  Shrikanth S. Narayanan,et al.  Toward detecting emotions in spoken dialogs , 2005, IEEE Transactions on Speech and Audio Processing.

[9]  Hermann Ney,et al.  Formant estimation for speech recognition , 1998, IEEE Trans. Speech Audio Process..

[10]  Mohamed S. Kamel,et al.  Segment-based approach to the recognition of emotions in speech , 2005, 2005 IEEE International Conference on Multimedia and Expo.

[11]  Shuichi Fukuda Detecting Emotions and Dangerous Actions for Better Human-System Team Working , 2008, 2008 Second International Conference on Secure System Integration and Reliability Improvement.

[12]  Raji Sukumar Discrete Wavelet Transforms and Artificial Neural Networks for Speech Emotion Recognition , 2010 .

[13]  Lijiang Chen,et al.  Speaker Independent Emotion Recognition Using HMMs Fusion System with Relative Features , 2008, 2008 First International Conference on Intelligent Networks and Intelligent Systems.

[14]  Patrick van der Smagt,et al.  Introduction to neural networks , 1995, The Lancet.

[15]  George N. Votsis,et al.  Emotion recognition in human-computer interaction , 2001, IEEE Signal Process. Mag..

[16]  A RajiSukumar.,et al.  Discrete Wavelet Transforms and Artificial Neural Networks for Speech Emotion Recognition , 2010 .

[17]  Astrid Paeschke,et al.  A database of German emotional speech , 2005, INTERSPEECH.

[18]  Lijiang Chen,et al.  Relative Speech Emotion Recognition Based Artificial Neural Network , 2008, 2008 IEEE Pacific-Asia Workshop on Computational Intelligence and Industrial Application.

[19]  B. Tomas,et al.  Influence of Emotions to pitch harmonics parameters of vowel /a/ , 2007, ELMAR 2007.

[20]  Richard M. Golden,et al.  Mathematical Methods for Neural Network Analysis and Design , 1996 .