Emotion detection using average relative amplitude features through speech

In this research work, a novel approach to emotion identification system is proposed for implementation in audio domain using human speech. In order to undertake the new approach, average relative bin frequency coefficients will be extracted from speech. In a noisy environment, audio data are not strictly aligned, thus getting proper noiseless signal is a challenge. Consequently, this affects the performance of emotion detection system. Due to these reasons, a newly proposed approach of Average Relative Bin Frequency technique in frequency domain will be implemented through audio data. Support vector machine with radial basis kernel will be used for the classification. Preliminary results showed an average of 86% accuracy for average relative frequency bin coefficients.

[1]  Léon J. M. Rothkrantz,et al.  Emotion Recognition from Speech by Combining Databases and Fusion of Classifiers , 2010, TSD.

[2]  George N. Votsis,et al.  Emotion recognition in human-computer interaction , 2001, IEEE Signal Process. Mag..

[3]  Nitin Thapliyal,et al.  Speech based Emotion Recognition with Gaussian Mixture Model , 2012 .

[4]  M. Y. Nayan,et al.  Emotion detection using relative amplitude-based features through speech , 2012, 2012 International Conference on Computer & Information Science (ICCIS).

[5]  Shashidhar G. Koolagudi,et al.  Emotion recognition from speech using source, system, and prosodic features , 2012, Int. J. Speech Technol..

[6]  Takashi Nose,et al.  A tone-modeling technique using a quantized F0 context to improve tone correctness in average-voice-based speech synthesis , 2012, Speech Commun..

[7]  Li Chen,et al.  Reliability detection by Fuzzy SVM with UBM Component feature for emotional speaker recognition , 2011, 2011 Eighth International Conference on Fuzzy Systems and Knowledge Discovery (FSKD).

[8]  Hichem Sahli,et al.  Relevance Vector Machine Based Speech Emotion Recognition , 2011, ACII.

[9]  Takashi Nose,et al.  HMM-Based Emphatic Speech Synthesis Using Unsupervised Context Labeling , 2011, INTERSPEECH.

[10]  Yixiong Pan,et al.  SPEECH EMOTION RECOGNITION USING SUPPORT VECTOR MACHINE , 2010 .

[11]  Björn W. Schuller,et al.  Recognising realistic emotions and affect in speech: State of the art and lessons learnt from the first challenge , 2011, Speech Commun..

[12]  Xin Wang,et al.  Facial expression anlysis using eye gaze information , 2011, 2011 IEEE International Conference on Computational Intelligence for Measurement Systems and Applications (CIMSA) Proceedings.

[13]  Jian Wang,et al.  Speech Emotion Recognition System Based on Integrating Feature and Improved HMM , 2012 .