Biodiversity Information Science and Standards : Conference Abstract
Conference Abstract
Towards the Acoustic Monitoring of Birds Migrating at Night
expand article infoHanna Pamula, Agnieszka Pocha§, Maciej Klaczynski
‡ AGH University of Science and Technology in Krakow, Faculty of Mechanical Engineering and Robotics, Department of Mechanics and Vibroacoustics, al. Mickiewicza 30 30-059, Krakow, Poland
§ Jagiellonian University, Faculty of Mathematics and Computer Science, Institute of Computer Science and Computational Mathematics, Department of Machine Learning, Łojasiewicza Street 6, 30-348, Krakow, Poland
Open Access


Every year billions of birds migrate between their breeding and wintering areas. As birds are an important indicator in nature conservation, migratory bird studies have been conducted for many decades, mostly by bird-ringing programmes and direct observation. However, most birds migrate at night, and therefore much information about their migration is lost. Novel methods have been developed to overcome this difficulty; including thermal imaging, radar, geolocation techniques, and acoustic recognition of bird calls.

Many bird species are detected by their characteristic sounds. This method of identification occurs more often than by direct observation, and therefore   recordings are widely used in avian research. The commonly used approach is to record the birds automatically, and to manually study the bird sounds in the recordings afterwards (Furnas and Callas 2015, Frommolt 2017). However, the tagging of recordings is a tedious and time-consuming process that requires expert knowledge, and, as a result, automatic detection of flight calls is in high demand. The first experiments towards this used energy thresholds or template matching (Bardeli et al. 2010, Towsey et al. 2012), and later on the machine and deep learning methods were applied (Stowell et al. 2018). Nevertheless, not many studies have focused specifically on night flight calls (Salamon et al. 2016, Lostanlen et al. 2018). Such acoustic monitoring could complement daytime avian research, especially when the field recording station is close to the bird-ringing station, as it is in our project.

In this study, we present the initial results of a long-term bird audio monitoring project using automatic methods for bird detection. Passive acoustic recorders were deployed at a narrow spit between a lake and the Baltic sea in Dąbkowice, West Pomeranian Voivodeship, Poland  . We recorded bird calls nightly from sunset till sunrise during the passerine autumn migration for 3 seasons. As a result, we collected over 3000 hours of recordings each season. We annotated a subset of over 50 hours, from different nights with various weather conditions. As avian flight calls are sporadic and short, we created a balanced set for training - recordings were divided into partially overlapping 500-ms clips, and we retained all clips containing calls and created about the same number of clips without bird sounds. Different signal representations were then examined (e.g. mel-spectrograms and multitaper). Afterwards, various convolutional neural networks were checked and their performance was compared using the area under the receiver operating characteristic curve (AUC) measure. Moreover, an initial attempt was made to take advantage of the transfer learning from image classification models. The results obtained by the deep learning methods are promising (AUC exceeding 80%), but higher bird detection accuracy is still needed. For a chosen bird species – Song thrush (Turdus philomelos) – we observed a correlation between calls recorded at night and birds caught in the nets during the day. This fact, as well as the promising results from the detection of calls from long-term recordings, indicate that acoustic monitoring of nocturnal birds has great potential and could be used to supplement the research of the phenomenon of seasonal bird migration.


acoustic monitoring, night flight calls, bird call detection, bioacoustics

Presenting author

Hanna Pamula

Presented at

Biodiversity_Next 2019 

Funding program

AGH University of Science and Technology, Grant Number:


login to comment