Deep Learning-Based Hazardous Sound Classification for the Hard of Hearing and Deaf

The hard of hearing or deaf can only access limited auditory information in dangerous situations. Therefore, development of a system for sensing hazardous auditory information may be of great help to them. However, such systems have focused on effective signal transduction when a hazardous sound is detected, and the classification of hazardous sounds has been less investigated. The present study was conducted to classify sounds by using Recurrent Neural Network (RNN)-based models, Convolutional Neural Network (CNN)-based models, the combination of the two models, and ensemble models prepared by combining various models. The experimental results showed that the accuracy of the 3-layer Long Short-Term Memory (LSTM) model was 97.63% and that of the ensemble model was 98.00%. As an attempt at real-life application of the developed model, a warning system was prepared by using Raspberry Pi and a vibrator.

[1]  Geoffrey E. Hinton,et al.  ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.

[2]  Rainer Brück,et al.  AUDIS wear: A smartwatch based assistive device for ubiquitous awareness of environmental sounds , 2016, 2016 38th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC).

[3]  Justin Salamon,et al.  A Dataset and Taxonomy for Urban Sound Research , 2014, ACM Multimedia.

[4]  Swetha Machanavajhala,et al.  A REAL-TIME ENVIRONMENTAL SOUND RECOGNITION SYSTEM FOR THE ANDROID OS , 2016 .

[5]  Jürgen Schmidhuber,et al.  Long Short-Term Memory , 1997, Neural Computation.

[6]  Thomas G. Dietterich Multiple Classifier Systems , 2000, Lecture Notes in Computer Science.

[7]  Sacha Krstulovic,et al.  Automatic Environmental Sound Recognition: Performance Versus Computational Cost , 2016, IEEE/ACM Transactions on Audio, Speech, and Language Processing.