Recognition of emotions in the Algerian Dialect Speech

Emotion recognition from speech has been an important and challenging research. The performance of speech features and influence of emotions number on the systems of emotion recognition in the Algerian dialect speech are studied in this work. To achieve this aim, an emotional database of the Algerian dialect (ADED) is built. This database contains four emotions: anger, fear, sadness and neutral. Extraction of features is an important step in the recognition of speech emotion. The features extracted in this study are: pitch, intensity, duration, unvoiced frames, jitter, shimmer, HNR (Harmonic Noise Ratio), formants and MFCCs (Mel Frequency Cepstral Coefficients). Each recognition system needs a classifier, so our system is based on the KNN (K-Nearest Neighbor) method of classification. Different results are obtained, the higher recognition rates are given when using a combination of features including pitch, intensity, duration, unvoiced frames, jitter, shimmer, HNR and the MFCCs parameters. And the performance is influenced by the number of emotions included: 82.29 % when using the four emotions of ADED database, 84.02% when three emotions fear, anger and neutral are recognized and 87.50% when only the fear and neutral emotions are used in the recognition system.

[1]  Kornel Laskowski,et al.  Emotion recognition in spontaneous speech using GMMs , 2006, INTERSPEECH.

[2]  Marc Schröder,et al.  Emotional speech synthesis: a review , 2001, INTERSPEECH.

[3]  Tsuyoshi Murata,et al.  {m , 1934, ACML.

[4]  Chao Yang,et al.  The fixed-point optimization of mel frequency cepstrum coefficients for speech recognition , 2011, Proceedings of 2011 6th International Forum on Strategic Technology.

[5]  Karima Meftouh,et al.  An Algerian dialect: Study and Resources , 2016, International Journal of Advanced Computer Science and Applications.

[6]  Koteswara Rao Anne,et al.  A comparative analysis of classifiers in emotion recognition through acoustic features , 2014, Int. J. Speech Technol..

[7]  Björn W. Schuller,et al.  Cross-language acoustic emotion recognition: An overview and some tendencies , 2015, 2015 International Conference on Affective Computing and Intelligent Interaction (ACII).

[8]  Noor Aina Zaidan,et al.  MFCC Global Features Selection in Improving Speech Emotion Recognition Rate , 2016, ICML 2016.

[9]  Sudarsana Reddy Kadiri,et al.  Neutral to anger speech conversion using non-uniform duration modification , 2014, 2014 9th International Conference on Industrial and Information Systems (ICIIS).

[10]  Theodoros Kostoulas,et al.  Enhancing Emotion Recognition from Speech through Feature Selection , 2010, TSD.

[11]  Gholamreza Anbarjafari,et al.  Efficiency of chosen speech descriptors in relation to emotion recognition , 2017, EURASIP Journal on Audio, Speech, and Music Processing.

[12]  Chung-Hsien Wu,et al.  Emotion Recognition of Affective Speech Based on Multiple Classifiers Using Acoustic-Prosodic Information and Semantic Labels , 2015, IEEE Transactions on Affective Computing.

[13]  Ling Guan,et al.  An investigation of speech-based human emotion recognition , 2004, IEEE 6th Workshop on Multimedia Signal Processing, 2004..

[14]  Mansour Sheikhan,et al.  Emotion recognition improvement using normalized formant supplementary features by hybrid of DTW-MLP-GMM model , 2012, Neural Computing and Applications.

[15]  Sachin Singh,et al.  A Survey on Autonomous Techniques for Music Classification based on Human Emotions Recognition , 2020 .

[16]  Ling Guan,et al.  A neural network approach for human emotion recognition in speech , 2004, 2004 IEEE International Symposium on Circuits and Systems (IEEE Cat. No.04CH37512).

[17]  Björn Schuller,et al.  Being bored? Recognising natural interest by extensive audiovisual integration for real-life application , 2009, Image Vis. Comput..

[18]  Björn W. Schuller,et al.  Recognising realistic emotions and affect in speech: State of the art and lessons learnt from the first challenge , 2011, Speech Commun..

[19]  Paul Boersma,et al.  Praat, a system for doing phonetics by computer , 2002 .

[20]  2021 International Seminar on Application for Technology of Information and Communication (iSemantic) , 2021 .

[21]  Adel M. Alimi,et al.  Speech emotion recognition based on Arabic features , 2015, 2015 15th International Conference on Intelligent Systems Design and Applications (ISDA).

[22]  Kamel Smaïli,et al.  Development of the Arabic Loria Automatic Speech Recognition system (ALASR) and its evaluation for Algerian dialect , 2017, ACLING.

[23]  Shashidhar G. Koolagudi,et al.  Identification of Hindi Dialects and Emotions using Spectral and Prosodic features of Speech , 2013 .

[24]  Lamiaa Abdel-Hamid,et al.  Egyptian Arabic speech emotion recognition using prosodic, spectral and wavelet features , 2020, Speech Commun..

[25]  T. Kishore Kumar,et al.  Emotion Recognition of Stressed Speech Using Teager Energy and Linear Prediction Features , 2018, 2018 IEEE 18th International Conference on Advanced Learning Technologies (ICALT).

[26]  Ragini Verma,et al.  Class-level spectral features for emotion recognition , 2010, Speech Commun..

[27]  andf , 2018, Numerical Methods for Engineers and Scientists.

[28]  Say Wei Foo,et al.  Classification of stress in speech using linear and nonlinear features , 2003, 2003 IEEE International Conference on Acoustics, Speech, and Signal Processing, 2003. Proceedings. (ICASSP '03)..

[29]  Ali Hassan On automatic emotion classification using acoustic features , 2012 .

[30]  Ziad Osman,et al.  Emotion recognition in Arabic speech , 2017, 2017 Sensors Networks Smart and Emerging Technologies (SENSET).

[31]  Kaya Oguz,et al.  Speech emotion recognition: Emotional models, databases, features, preprocessing methods, supporting modalities, and classifiers , 2020, Speech Commun..

[32]  Danna Zhou,et al.  d. , 1840, Microbial pathogenesis.

[33]  Semiye Demircan,et al.  Application of fuzzy C-means clustering algorithm to spectral features for emotion classification from speech , 2016, Neural Computing and Applications.

[34]  Jon Sánchez,et al.  Automatic emotion recognition using prosodic parameters , 2005, INTERSPEECH.

[35]  Alex Pentland,et al.  Human computing and machine understanding of human behavior: a survey , 2006, ICMI '06.

[36]  Wojciech Majewski,et al.  Polish Emotional Speech Database - Recording and Preliminary Validation , 2009, COST 2102 Conference.

[37]  Abdelmajid Farchi,et al.  Emotion recognition from Moroccan dialect speech and energy band distribution , 2019, 2019 International Conference on Wireless Technologies, Embedded and Intelligent Systems (WITS).

[38]  Hadeel Al-Obaidy,et al.  Predicting Software Faults Based on K-Nearest Neighbors Classification , 2019 .

[39]  P. Aruna,et al.  Applying Machine Learning Techniques for Speech Emotion Recognition , 2018, 2018 9th International Conference on Computing, Communication and Networking Technologies (ICCCNT).

[40]  W. Marsden I and J , 2012 .

[41]  James D. Edge,et al.  Audio-visual feature selection and reduction for emotion classification , 2008, AVSP.

[42]  Ibrahiem M. M. El Emary,et al.  Speech emotion recognition approaches in human computer interaction , 2013, Telecommun. Syst..

[43]  Francesc Alías,et al.  The role of prosody and voice quality in indirect storytelling speech: A cross-narrator perspective in four European languages , 2017, Speech Commun..

[44]  Muljono,et al.  Comparison Method in Indonesian Emotion Speech Classification , 2019, 2019 International Seminar on Application for Technology of Information and Communication (iSemantic).

[45]  Shashidhar G. Koolagudi,et al.  Choice of a classifier, based on properties of a dataset: case study-speech emotion recognition , 2018, International Journal of Speech Technology.

[46]  Agnes Jacob Speech emotion recognition based on minimal voice quality features , 2016, 2016 International Conference on Communication and Signal Processing (ICCSP).

[47]  G. Sivaranjani,et al.  EMOTION RECOGNITION FROM SPEECH WITH GAUSSIAN MIXTURE MODELS AND VIA BOOSTED GMM , 2018 .

[48]  Shrikanth Narayanan,et al.  Recognition of negative emotions from the speech signal , 2001, IEEE Workshop on Automatic Speech Recognition and Understanding, 2001. ASRU '01..

[49]  Werner Verhelst,et al.  An evaluation of the robustness of existing supervised machine learning approaches to the classification of emotions in speech , 2007, Speech Commun..

[50]  Aurobinda Routray,et al.  Study of prosodic feature extraction for multidialectal Odia speech emotion recognition , 2016, 2016 IEEE Region 10 Conference (TENCON).

[51]  Cynthia Breazeal,et al.  Recognition of Affective Communicative Intent in Robot-Directed Speech , 2002, Auton. Robots.

[52]  F. A.,et al.  Automatic Emotion Recognition from Speech Using Artificial Neural Networks with Gender-Dependent Databases , 2009, 2009 International Conference on Advances in Computing, Control, and Telecommunication Technologies.

[53]  Elliot Moore,et al.  Critical Analysis of the Impact of Glottal Features in the Classification of Clinical Depression in Speech , 2008, IEEE Transactions on Biomedical Engineering.

[54]  Mohamed Mbarki,et al.  Automatic speech emotion recognition using an optimal combination of features based on EMD-TKEO , 2019, Speech Commun..

[55]  Ismail Shahin,et al.  Emotion Recognition Using Hybrid Gaussian Mixture Model and Deep Neural Network , 2019, IEEE Access.

[56]  Turgut Özseven,et al.  A novel feature selection method for speech emotion recognition , 2019, Applied Acoustics.

[57]  Yongzhao Zhan,et al.  Adaptive and Optimal Classification of Speech Emotion Recognition , 2008, 2008 Fourth International Conference on Natural Computation.

[58]  Sartra Wongthanavasu,et al.  Speech emotion recognition using Support Vector Machines , 2013, 2013 5th International Conference on Knowledge and Smart Technology (KST).

[59]  Nilesh R. Patel,et al.  Implementation and Comparison of Speech Emotion Recognition System Using Gaussian Mixture Model (GMM) and K- Nearest Neighbor (K-NN) Techniques , 2015 .

[60]  Dipankar Das,et al.  Identifying Psychological Theme Words from Emotion Annotated Interviews , 2013, MIKE.

[61]  Loïc Kessous,et al.  The relevance of feature type for the automatic classification of emotional user states: low level descriptors and functionals , 2007, INTERSPEECH.

[62]  Glenn I. Roisman,et al.  The emotional integration of childhood experience: physiological, facial expressive, and self-reported emotional response during the adult attachment interview. , 2004, Developmental psychology.