=Paper=
{{Paper
|id=Vol-2125/paper_85
|storemode=property
|title=A Baseline for Large-Scale Bird Species Identification in Field Recordings
|pdfUrl=https://ceur-ws.org/Vol-2125/paper_85.pdf
|volume=Vol-2125
|authors=Stefan Kahl,Thomas Wilhelm-Stein,Holger Klinck,Danny Kowerko,Maximilian Eibl
|dblpUrl=https://dblp.org/rec/conf/clef/KahlWKKE18
}}
==A Baseline for Large-Scale Bird Species Identification in Field Recordings==
A Baseline for Large-Scale Bird Species Identification in Field Recordings Stefan Kahl1 , Thomas Wilhelm-Stein1 , Holger Klinck2 , Danny Kowerko3 , and Maximilian Eibl1 1 Chair Media Informatics, Chemnitz University of Technology, D-09107 Chemnitz, Germany 2 Junior Professorship Media Computing, Chemnitz University of Technology, D-09107 Chemnitz, Germany 3 Bioacoustics Research Program, Cornell Lab of Ornithology, 159 Sapsucker Woods Road, Ithaca, NY 14850, USA {stefan.kahl, thomas.wilhelm-stein, danny.kowerko, maximilian.eibl}@informatik.tu-chemnitz.de, Holger.Klinck@cornell.edu Abstract. The LifeCLEF bird identifcation task poses a difficult chal- lenge in the domain of acoustic event classification. Deep learning tech- niques have greatly impacted the field of bird sound recognition in recent years. We discuss our attempt of large-scale bird species identification using the 2018 BirdCLEF baseline system. Keywords: Bioacoustics · Bird Sounds · Deep Learning · BirdCLEF. 1 Motivation Large-scale bird sound identification in audio recordings is the foundation of long-term species diversity monitoring. Aiding this labor intensive task with automated systems that can recognize multiple hundreds of species has been the focus in recent years. As part of the 2018 LifeCLEF workshop [1], the BirdCLEF bird identification challenges [2] provide large datasets containing almost 50.000 recordings to assess the performance of various systems attempting to push the boundaries of automated bird sound recognition. 2 Related Work In 2016, Sprengel et al. [3] demonstrated the superior performance of convo- lutional neural networks (CNN) for the classification of bird sounds. Following that approach, we were able to improve the performance on a larger dataset con- taining 1500 different species with our 2017 BirdCLEF participation [4]. This year, we present an implementation of a streamlined workflow built on the most fundamental principles of visual classification using CNN. We published the code repository as baseline system complementing the 2018 BirdCLEF challenge [5]. The following workflow design, training scheme and submission results are en- tirely based on that system, establishing a good overall baseline for future com- parisons and improvements. 3 Workflow The key stages of our workflow include dataset pre-processing, spectrogram ex- traction, CNN training and evaluation. We adopted our last year’s attempt and focused mainly on basic deep learning techniques, keeping the code base as simple and comprehensible as possible, while maintaining a good overall performance. 3.1 Dataset Handling Using convolutional neural networks for the classification of acoustic events proved to be very effective despite the fact that these techniques are tailor made for visual recognition. Representing audio recordings as spectrograms overcomes this gap between the two domains of audio and image. We decided to use MEL- scale log-amplitude spectrograms which have been effectively used in similar approaches (e.g. [6]). A more detailed description of the extraction and pre- processing process can be found in [5]. 3.2 Training Our baseline training process supports multiple shallow and deep model archi- tectures, extensive dataset augmentation, learning rate scheduling, model pre- training and result pooling. We implemented two basic CNN concepts: Fully- convolutional architectures with simple layer sequences and ResNet variations with shortcut connections. We also provide eBird4 checklist metadata for both soundscape locations in Peru and Columbia along with the basline repository. 3.3 Model Distillation Most CNN implementations are computationally expensive and rely on power- hungry hardware. Future applications of automated bird sound recognition will include field recorders capable of not only of recording, but also analyzing audio data in real-time. In those cases, battery life becomes an issue. In recent years, (semi-) mobile hardware - mostly used for IoT-applications - has been designed to aid this task. However, those hardware platforms are not yet suited for deep learning inference using complex models. In 2015, Hinton et. al [7] presented an approach to distill knowledge in neural networks. We followed that scheme of model distillation and implemented a basic variant of teacher-student learning. Our baseline system allows to replace binary training targets with log-probability predictions of either single models or entire ensembles. We designed a simple shallow model that can predict species probabilities of one-second audio chunks in less than one second running on a Raspberry Pi 3+. The resulting scores are slightly lower than those of large single models, but still above the initial capabilities of the tiny CNN model. 4 www.ebird.org/explore The prediction performance of this approach is promising and model distillation may have significant impact on the field of mobile real-time species diversity assessment. 4 Results We tried to cover different basic training and prediction schemes with our run submissions, including single baseline models, large and diverse model ensembles, metadata assisted attempts with species pre-selection and knowledge distillation training of tiny models. Table 1 provides an overview of selected results from our submissions. Table 1. Selected submission results (run IDs in brackets, not all runs listed for clar- ity). Large ensembles including different net architectures and dataset splits perform best. Pre-selecting species for specific locations does not improve the results. Model distillation helps to reduce computational costs and can maintain results that are com- parable with the performance of large single nets. Monophone Task Soundscape Task Run MRR MRR c-mAP c-mAP Description Foreground Background Peru Columbia Best Single 0.487 (1) 0.448 (1) - - Best Ensemble 0.644 (5) 0.588 (5) 0.086 (6) 0.117 (6) Pre-Selection - - 0.081 (7) 0.052 (8) Raspberry Pi 0.425 (4) 0.385 (4) 0.077 (9) 0.083 (9) The results show that our baseline attempt yields competitive results con- sidering the complex evaluation task. Most results did match our expectations for the audio-only classification of field recordings. The key takeaways of the analysis of the submission results are: – Diverse model ensembles covering different net architectures and dataset splits outperform single neural nets by a significant margin. This comes as no surprise in the domain of metric-centered competitions, but might not be applicable to real-world scenarios due to increased computational costs. – Pre-selecting species did not improve the overall performance as expected. In some cases, selecting species based on time of the year and location helps to reduce training time. Using metadata as post-filter to eliminate false de- tections or as input during model training might lead to better results. – Model distillation is a powerful tool to increase the classification perfor- mance of tiny neural networks. The results show comparable performance in the soundscape domain despite much smaller model architectures, when compared to model ensembles. We published our entire code repository5 and encourage future participants and interested research groups to build upon our results and improve the per- formance for the analysis of complex soundscapes - the most crucial aspect of species diversity monitoring. 5 Future Work Assessing high-quality field recordings for the presence of bird species using con- volutional neural networks is an effective application of deep learning techniques to the domain of acoustic event detection. Considering our own scores and those of other participants, current machine learning algorithms yield very strong re- sults for this task. However, the 2018 BirdCLEF evaluation showed that the transfer of knowledge extracted from monophonic community recordings to the domain of long-term soundscape recordings is still very difficult. Hardly any improvements over last year’s result have been accomplished. Future research should specifically focus on this task. Additionally, power-hungry hardware and computationally expensive algorithms are not well-suited for real-world applica- tions such as mobile reorders. Improving techniques to shrink the size of neural networks while maintaining the overall performance will greatly help the field of long-term species diversity assessment. References 1. Joly, A., Goëau, H., Botella, C., Glotin, H., Bonnet, P., Planqué, R., Vellinga, W.-P., Müller, H.: Overview of LifeCLEF 2018: a large-scale evaluation of species identifi- cation and recommendation algorithms in the era of AI. In: Proceedings of CLEF 2018 (2018). 2. Goëau, H., Glotin, H., Planqué, R., Vellinga, W.-P., Kahl, S., Joly, A.: Overview of BirdCLEF 2018: monophone vs. soundscape bird identification. In: CLEF working notes 2018 (2018). 3. Sprengel, E., Martin Jaggi, Y. K., Hofmann, T.: Audio based bird species identifi- cation using deep learning techniques. In: Working notes of CLEF 2016 (2016). 4. Kahl, S., Wilhelm-Stein, T., Hussein, H., Klinck, H., Kowerko, D., Ritter, M., Eibl, M.: Large-Scale Bird Sound Classification using Convolutional Neural Networks. In: CLEF working notes 2017 (2017). 5. Kahl, S., Wilhelm-Stein, T., Klinck, H., Kowerko, D., Eibl, M.: Recognizing Birds from Sound - The 2018 BirdCLEF Baseline System. arXiv preprint arXiv:1804.07177 (2018). 6. Grill, T., Schlüter, J.: Two convolutional neural networks for bird detection in audio signals. In: Signal Processing Conference (EUSIPCO), 2017 25th European, pp. 1764–1768, IEEE (2017). 7. Hinton, G., Vinyals, O., Dean, J.: Distilling the knowledge in a neural network. arXiv preprint arXiv:1503.02531 (2015). 5 https://github.com/kahst/BirdCLEF-Baseline