Deep Learning Locally Trained Wildlife Sensing in Real Acoustic Wetland Environment

We describe ‘Tidzam’, an application of deep learning that leverages a dense, multimodal sensor network installed at a large wetland restoration performed at Tidmarsh, a 600-acre former industrial-scale cranberry farm in Southern Massachusetts. Wildlife acoustic monitoring is a crucial metric during post-restoration evaluation of the processes, as well as a challenge in such a noisy outdoor environment. This article presents the entire Tidzam system, which has been designed in order to identify in real-time the ambient sounds of weather conditions as well as sonic events such as insects, small animals and local bird species from microphones deployed on the site. This experiment provides insight on the usage of deep learning technology in a real deployment. The originality of this work concerns the system’s ability to construct its own database from local audio sampling under the supervision of human visitors and bird experts.

[1]  Luis J. Villanueva-Rivera,et al.  Using Automated Digital Recording Systems as Effective Tools for the Monitoring of Birds and Amphibians , 2006 .

[2]  Florian Metze,et al.  A comparison of Deep Learning methods for environmental sound detection , 2017, 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).

[3]  Qiang Huang,et al.  Unsupervised Feature Learning Based on Deep Models for Environmental Audio Tagging , 2016, IEEE/ACM Transactions on Audio, Speech, and Language Processing.

[4]  Tuomas Virtanen,et al.  Stacked convolutional and recurrent neural networks for bird audio detection , 2017, 2017 25th European Signal Processing Conference (EUSIPCO).

[5]  Aren Jansen,et al.  CNN architectures for large-scale audio classification , 2016, 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).

[6]  Ieee Staff 2017 25th European Signal Processing Conference (EUSIPCO) , 2017 .

[7]  Kazuhiro Nakadai,et al.  Bird Song Scene Analysis Using a Spatial-Cue-Based Probabilistic Model , 2017, J. Robotics Mechatronics.

[8]  Jill L. Deppe,et al.  Using soundscape recordings to estimate bird species abundance, richness, and composition , 2009 .

[9]  Joseph A. Paradiso,et al.  The Networked Sensory Landscape: Capturing and Experiencing Ecological Change Across Scales , 2017, PRESENCE: Teleoperators and Virtual Environments.

[10]  Tuomas Virtanen,et al.  Convolutional recurrent neural networks for bird audio detection , 2017, 2017 25th European Signal Processing Conference (EUSIPCO).