Acoustics Speech Processing of Sanskrit Language

Speech processing (SP) is the latest trend in technology. An intelligent and precise human-machine interaction (HMI) is designed to engineer an automated, smart and secure application for household and commercial application. The existing methods highlight the absence of the speech processing in the under-resourced languages. The novelty of this work is that it presents a study of acoustic speech processing (ASP) using spectral components of Mel frequency cepstrum coefficient (MFCC) of Sanskrit language. A customized speech database is created as no generic database is available in Sanskrit. The processing method includes speech signal isolation, feature selection and extraction of selected features for applications. The speech is processed over a custom dataset consisting of Sanskrit speech corpus. The spectral features are calculated over 13 coefficients providing improved performance. The results obtained highlight the performance of the proposed system with the variation of the lifter parameter.

[1]  Nitin Trivedi,et al.  Speech Recognition by Wavelet Analysis , 2011 .

[2]  Ajai Jain,et al.  Architecture of English to Sanskrit machine translation , 2015, 2015 SAI Intelligent Systems Conference (IntelliSys).

[3]  Yuan-Ting Zhang,et al.  Bionic wavelet transform: a new time-frequency method based on an auditory model , 2001, IEEE Trans. Biomed. Eng..

[4]  Sahar Kalid Ahmed,et al.  Emotion Recognition in Speech Using Neural Network , 2008 .

[5]  Soumya Sahoo,et al.  Word Extraction from Speech Recognition using Correlation Coefficients , 2012 .

[6]  S. B. Dhonde,et al.  Significance of Frequency Band Selection of MFCC for Text-Independent Speaker Identification , 2017 .

[7]  Alessandra Russo,et al.  Speech Emotion Classification Using Machine Learning Algorithms , 2008, 2008 IEEE International Conference on Semantic Computing.

[8]  Nursuriati Jamil,et al.  Influences of age in emotion recognition of spontaneous speech: A case of an under-resourced language , 2017, 2017 International Conference on Speech Technology and Human-Computer Dialogue (SpeD).

[9]  Rishi Pal Singh,et al.  Automatic Speech Recognition: A Review , 2012 .

[10]  Azam Bastanfard,et al.  Real-time speech emotion recognition by minimum number of features , 2016, 2016 Artificial Intelligence and Robotics (IRANOPEN).

[11]  Zhaoyan Zhang Mechanics of human voice production and control. , 2016, The Journal of the Acoustical Society of America.

[12]  Ramesh Gamasu ECG Based Integrated Mobile Tele medicine System for Emergency Health Tribulations , 2014, BSBT 2014.

[13]  Tiago H. Falk,et al.  Automatic speech emotion recognition using modulation spectral features , 2011, Speech Commun..

[14]  Pierre-Yves Oudeyer,et al.  The production and recognition of emotions in speech: features and algorithms , 2003, Int. J. Hum. Comput. Stud..

[15]  Moirangthem Tiken Singh,et al.  A Survey Report on Speech Recognition System , 2015 .

[16]  Wolfgang Wahlster,et al.  Verbmobil: Foundations of Speech-to-Speech Translation , 2000, Artificial Intelligence.

[17]  Kiavash Bahreini,et al.  Towards real-time speech emotion recognition for affective e-learning , 2015, Education and Information Technologies.

[18]  Ahmed Hammouch,et al.  Detecting Patients with Parkinson ’ s disease using Mel Frequency Cepstral Coefficients and Support Vector Machines , 2015 .

[19]  John H. L. Hansen,et al.  A Review on Speech Recognition Technique , 2010 .

[20]  Sandeep Sharma,et al.  Speech Recognition System: A Review , 2015 .

[21]  Ayushi Y. Vadwala,et al.  Survey paper on Different Speech Recognition Algorithm: Challenges and Techniques , 2017 .

[22]  Ajay Anil Gurjar,et al.  Analysis and Dissection of Sanskrit Divine Sound "OM" Using Digital Signal Processing to Study the Science Behind "OM" Chanting , 2016, 2016 7th International Conference on Intelligent Systems, Modelling and Simulation (ISMS).

[23]  Theodoros Iliou,et al.  Features and classifiers for emotion recognition from speech: a survey from 2000 to 2011 , 2012, Artificial Intelligence Review.

[24]  Maulin Joshi,et al.  Speaker Recognition Using MFCC and Hybrid Model of VQ and GMM , 2013, ISI.

[25]  Shashidhar G. Koolagudi,et al.  Emotion recognition from speech: a review , 2012, International Journal of Speech Technology.

[26]  C. Pierrakeas,et al.  Online Collaboration Environments in Telemedicine Applications of Speech Therapy , 2005, 2005 IEEE Engineering in Medicine and Biology 27th Annual Conference.

[27]  Rajesvary Rajoo,et al.  Influences of languages in speech emotion recognition: A comparative study using Malay, English and Mandarin languages , 2016, 2016 IEEE Symposium on Computer Applications & Industrial Electronics (ISCAIE).

[28]  Gholamreza Anbarjafari,et al.  Speech-based emotion recognition and next reaction prediction , 2017, 2017 25th Signal Processing and Communications Applications Conference (SIU).

[29]  B. Moore The Role of Temporal Fine Structure Processing in Pitch Perception, Masking, and Speech Perception for Normal-Hearing and Hearing-Impaired People , 2008, Journal of the Association for Research in Otolaryngology.