Efficient Feature Extraction for Fear State Analysis from Human Voice

Background/Objectives: Analysis of human speech emotion has been continued since long. As the study and recognition helps the society in many respects, we intend to analyze the similar type of emotions. Methods/Statistical Analysis: ‘Fear’ and ‘Nervousness’ are being analyzed in comparison with normal voice. The correlation between these two emotions found to be very close. These voices belong to Oriya language. The popular features of speech, Mel-frequency cepstral coefficients (MFCCs) are used. As the fundamental frequency is unique from voice to voice, it is a suitable feature in case of similar voice signals. Findings: The combination of these two features outperformed the single feature based classification. In addition, the performance has been measured using log-likelihood ratio parameter. For recognition purpose, Gaussian mixture model (GMM) has been selected, and tested for these features. Novelty/Improvement: The individual MFCCs show 81.33%, whereas the combined features show 86.01% of accuracy. It is clearly evidenced in the result section.

[1]  G. Gunasekaran,et al.  Fuzzy Logic based Nam Speech Recognition for Tamil Syllables , 2016 .

[2]  Mihir Narayan Mohanty,et al.  Classification of Emotions of Angry and Disgust , 2015, Smart Comput. Rev..

[3]  Latifah Ananda Chaeron Detecting the emotion fear through voice Version of June 25 , 2008 .

[4]  N. Rajkumar,et al.  Cognitive Intelligent Tutoring System based on Affective State , 2015 .

[5]  Lawrence R. Rabiner,et al.  On the use of autocorrelation analysis for pitch detection , 1977 .

[6]  Fakhri Karray,et al.  Survey on speech emotion recognition: Features, classification schemes, and databases , 2011, Pattern Recognit..

[7]  Chung-Hsien Wu,et al.  Speech Emotion Verification Using Emotion Variance Modeling and Discriminant Scale-Frequency Maps , 2015, IEEE/ACM Transactions on Audio, Speech, and Language Processing.

[8]  Mihir Narayan Mohanty,et al.  Efficient feature combination techniques for emotional speech classification , 2016, International Journal of Speech Technology.

[9]  G. N. Rathna,et al.  Speech Emotion Recognition: Performance Analysis based on Fused Algorithms and GMM Modelling , 2016 .

[10]  Jianwu Dang,et al.  Physiological Feature Extraction for Text Independent Speaker Identification using Non-Uniform Subband Processing , 2007, 2007 IEEE International Conference on Acoustics, Speech and Signal Processing - ICASSP '07.

[11]  John H. L. Hansen,et al.  Nonlinear analysis and classification of speech under stressed conditions , 1994 .

[12]  Tiago H. Falk,et al.  Automatic speech emotion recognition using modulation spectral features , 2011, Speech Commun..

[13]  Daniel McDuff,et al.  AffectAura: Emotional wellbeing reflection system , 2012, 2012 6th International Conference on Pervasive Computing Technologies for Healthcare (PervasiveHealth) and Workshops.

[14]  Zhigang Deng,et al.  An acoustic study of emotions expressed in speech , 2004, INTERSPEECH.

[15]  Swati Vaid,et al.  Classification of Human Emotions using Multiwavelet Transform based Features and Random Forest Technique , 2015 .

[16]  Mihir Narayan Mohanty,et al.  New features for emotional speech recognition , 2015, 2015 IEEE Power, Communication and Information Technology Conference (PCITC).

[17]  Mihir Narayan Mohanty,et al.  Classification of Emotional Speech of Children Using Probabilistic Neural Network , 2015 .

[18]  Tanja Neumann Introduction To The Theory Of Statistics , 2016 .

[19]  L. Devillers,et al.  F0 and pause features analysis for Anger and Fear detection in real-life spoken dialogs , 2004 .

[20]  Robert L. Wolpert,et al.  Statistical Inference , 2019, Encyclopedia of Social Network Analysis and Mining.

[21]  Chloé Clavel,et al.  Fear-type emotion recognition for future audio-based surveillance systems , 2008, Speech Commun..

[22]  N. Venkatram,et al.  Establishing Efficient Security Scheme in Home IOT Devices through Biometric Finger Print Technique , 2016 .

[23]  Constantine Kotropoulos,et al.  Emotional speech recognition: Resources, features, and methods , 2006, Speech Commun..

[24]  Pawan Kumar,et al.  Pitch-based cepstral features for gender classification in noisy environments , 2013 .

[25]  Ning An,et al.  Speech Emotion Recognition Using Fourier Parameters , 2015, IEEE Transactions on Affective Computing.

[26]  Mihir Narayan Mohanty,et al.  Analysis of stressed human speech , 2011, Int. J. Comput. Vis. Robotics.