=Paper=
{{Paper
|id=Vol-3740/paper-202
|storemode=property
|title=Transfer Learning with Pseudo Multi-Label Birdcall Classification for DS@GT BirdCLEF
2024
|pdfUrl=https://ceur-ws.org/Vol-3740/paper-202.pdf
|volume=Vol-3740
|authors=Anthony Miyaguchi,Adrian Cheung,Murilo Gustineli,Ashley Kim
|dblpUrl=https://dblp.org/rec/conf/clef/MiyaguchiCGK24
}}
==Transfer Learning with Pseudo Multi-Label Birdcall Classification for DS@GT BirdCLEF
2024==
Transfer Learning with Pseudo Multi-Label Birdcall
Classification for DS@GT BirdCLEF 2024
Anthony Miyaguchi1,* , Adrian Cheung1,* , Murilo Gustineli1,* and Ashley Kim1
1
Georgia Institute of Technology, North Ave NW, Atlanta, GA 30332
Abstract
We present working notes for the DS@GT team on transfer learning with pseudo multi-label birdcall clas-
sification for the BirdCLEF 2024 competition, focused on identifying Indian bird species in recorded sound-
scapes. Our approach utilizes production-grade models such as the Google Bird Vocalization Classifier, Bird-
NET, and EnCodec to address representation and labeling challenges in the competition. We explore the dis-
tributional shift between this year’s edition of unlabeled soundscapes representative of the hidden test set
and propose a pseudo multi-label classification strategy to leverage the unlabeled data. Our highest post-
competition public leaderboard score is 0.63 using BirdNET embeddings with Bird Vocalization pseudo-labels.
Our code is available at github.com/dsgt-kaggle-clef/birdclef-2024.
Keywords
Transfer Learning, Dataset Annotation, Embeddings, Association Rule Mining, Google Bird Vocalization Classifier,
BirdNET, EnCodec, CEUR-WS
1. Introduction
The BirdCLEF 2024 competition [1] involves identifying bird species in 4-minute-long soundscapes
recorded in the Western Ghats of India as part of the LifeCLEF lab [2]. Passive monitoring of ecological
areas allows humans to determine how to allocate our attention to preserve biodiversity for posterity.
The primary method of passive monitoring for birds is through bioacoustics. Given a recording of a
soundscape located on an autonomous recording device in the field, we would like to determine when
and where specific birds are vocalizing.
The objective of the BirdCLEF competition is to predict the presence of each of the 182 target species
for every 5-second segment in test soundscapes. One of the main challenges this year is the 4400 minutes
of test soundscapes that must be predicted in 120 minutes of CPU time instead of 2000 minutes in
BirdCLEF 2023 [3]. We focus on transfer learning using Google’s Bird Vocalization Classification Model,
made publicly available in the BirdCLEF 2023 competition. We explore embeddings from BirdNET, a
model trained on bird vocalizations, and EnCodec, a neural audio codec trained on diverse audio data.
2. Birdcall Classification Overview
Birdcall classification is a challenging task due to the variability in bird vocalizations, the presence of
background noise, and the large number of species to classify. In addition, the measured data comes in
the form of audio recordings, which are high-dimensional and require specialized processing techniques.
Many successful approaches to birdcall classification utilize convolutional neural networks (CNNs) to
extract features from audio spectrograms. Audio spectrograms are a time-frequency representation
of the audio signal, which are extracted using the short-time Fourier transform (STFT) [4], often with
CLEF 2024: Conference and Labs of the Evaluation Forum, September 9-12, 2024, Grenoble, France
*
Corresponding author.
$ acmiyaguchi@gatech.edu (A. Miyaguchi); acheung@gatech.edu (A. Cheung); murilogustineli@gatech.edu
(M. Gustineli); akim614@gatech.edu (A. Kim)
https://linkedin.com/in/acmiyaguchi (A. Miyaguchi); https://linkedin.com/in/acheunggt (A. Cheung);
https://linkedin.com/in/murilo-gustineli (M. Gustineli); https://www.linkedin.com/in/-ashleykim/ (A. Kim)
0000-0002-9165-8718 (A. Miyaguchi); 0009-0006-8650-4550 (A. Cheung); 0009-0003-9818-496X (M. Gustineli)
© 2024 Copyright for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0).
CEUR
ceur-ws.org
Workshop ISSN 1613-0073
Proceedings
additional preprocessing steps such as mel-frequency scaling. The spectrograms are represented as 2D
images, which utilize techniques in the rich literature surrounding image classification.
BirdNET is a popular birdcall classification model that utilizes the spectrogram-CNN approach. It
is widely distributed in the field due to its high accuracy and ease of use on mobile devices [5]. The
Google Bird Vocalization Classifier is another model using EfficientNet-B1, a similar CNN architecture,
and is trained on many soundscapes. It was released alongside the BirdCLEF 2023 competition and has
more than 10,000 species in its output space [6].
3. Domain Knowledge Transfer via Embedding Spaces
Neural networks are universal function approximators that map some input space to an arbitrary output
space [7]. A neural network’s intermediate layers can be considered a manifold that typically projects
high dimensional data to lower dimensional spaces. Transfer learning is a technique that leverages the
learned representations of a model trained on one task to improve the performance of a model on a
different task. In the context of birdcall classifications, raw audio data is projected onto a manifold
optimized to discriminate between bird species. The projections are called embeddings and can be used
as features in downstream tasks to transfer knowledge to a new but similar domain. Few-shot learning
on global bird embeddings tends to be effective on new domains as per Ghani et al. [8].
Figure 1: PaCMAP projections of the top five species averaged embeddings ranked by soundscape fre-
quency. Embeddings can be evaluated qualitatively by clustering behavior.
We visually inspect learned embeddings by projecting them into two or three dimensions to reveal
meaningful semantic structures that can be qualitatively interpreted by humans. We use PaCMAP, a
dimensionality reduction technique that preserves local and global structure on a manifold via pairwise
relationships [9], to visualize embeddings from the Bird Vocalization Classifier, BirdNET, and EnCodec.
In Figure 1, we project the top five species by frequency in the soundscape dataset. The first two are
domain-specific models trained on bird vocalizations, while the latter is a general-purpose neural audio
codec. EnCodec is particularly interesting because it is not trained on bird vocalizations but rather a
diverse set of audio data using a self-supervised learning objective guided by self-attention mechanisms
[10]. It is also perceptually optimized for audio compression, preserving meaningful structures in the
embedding space, such as bird vocalizations over static background noise.
We explore the effectiveness of transfer learning, where bird vocalization predictions are surrogates for
true labels. We hypothesize that transfer learning is an effective technique for the competition because
existing models capture underlying structures amenable to optimization by simple linear classifiers. We
quantify how well we can learn domain-specific adaptations between different embedding spaces and
how well each of these models is suited to capture the underlying structure of the data.
4. Exploratory Data Analysis
We perform an exploratory analysis on the training and unlabeled soundscape datasets to understand
species distribution and their co-occurrence patterns. We hypothesize a shift in species distribution
between the training and unlabeled soundscape datasets due to the differences in the recording en-
vironments observed in downstream domain knowledge transfer. The training dataset contains 182
species obtained from crowd-sourced Xeno-Canto recordings. Because they are crowd-sourced, the
training data is likely biased toward clear and distinct vocalizations that typically occur in isolation. The
soundscape dataset is a collection of 4-minute soundscapes from Western Ghats, India, representative
of the hidden test set in the competition [11]. The soundscape is likely to be more intermittent and
contain vocalizations that are less distinct and overlapping due to the lack of human-directed attention
to the recording process.
Figure 2: Representation of the distribution of species detected in the train and unlabeled soundscapes,
sorted by the frequency of species in the soundscape.
We use the Bird Vocalization model to extract the embeddings and logits from the datasets in 5-
second intervals. The training dataset comprises 217k discrete intervals totaling 302 hours, while the
soundscape dataset has 407k intervals totaling 566 hours. We assume that an interval contains a call
if the maximum logit value run through a sigmoid function exceeds a threshold of 0.5. The training
dataset has a higher density of calls, with 62% of intervals containing at least one call, compared to
8.8% in the soundscape dataset.
We compute the relative frequency of species overall discrete intervals and compare their distributions
by the ranked frequency in the soundscape dataset in Figure 2. There is a notable discordance between
the two distributions, with many species in the training set unrepresented in the unlabeled soundscapes.
We observe that top species in each dataset do not align in Table 1 using the raw frequency of occurrences.
Table 1 Table 2
Top species for train and soundscape. Top frequent itemsets in soundscape.
Train Freq Soundscape Freq Items Freq
1 bncwoo3 14170 grnsan 6272 [comior1] 5320
2 grnsan 8069 comior1 5648 [lirplo] 5308
3 inbrob1 7189 lirplo 5638 [lirplo, comior1] 5307
4 comior1 6575 bkrfla1 5532 [bkrfla1] 5283
5 lirplo 6423 comtai1 5323 [bkrfla1, lirplo] 5280
6 bkrfla1 6260 btbeat1 5189 [bkrfla1, lirplo, comior1] 5280
7 comtai1 5664 putbab1 5009 [bkrfla1, comior1] 5280
8 houcro1 4774 purher1 4735 [grnsan] 5197
9 comsan 4214 whcbar1 4057 [grnsan, comior1] 5169
10 btbeat1 3749 mawthr1 3114 [grnsan, lirplo] 5168
Figure 3: The plot shows the distribution of itemset sizes in the training and soundscape datasets. The
distribution represents how likely species are to co-occur in each recording.
We use a frequent-pattern mining algorithm, FPGrowth [12], to identify co-occurrence patterns in
the soundscape dataset. In Table 2, we observe that co-occurrences of species can appear more often
than individual species alone. The frequent itemsets give us a rough estimate of how many birds we
can expect to see in a single recording. In Figure 3, we plot the distribution of normalized itemset sizes
in the training and soundscape datasets. We observe an approximately normal distribution of sizes
centered around four to six species per recording. The training set is skewed toward smaller itemsets,
likely due to biases in the data collection process where individuals are more likely to record and upload
isolated calls.
5. Methodology
The experiments are run over several modular stages. We implement an end-to-end workflow that
applies domain-specific fine-tuning to state-of-the-art models for birdcall classification. We quantify
differences between choices of dataset, architecture, and training losses. The second part of the
experiment focuses on transfer learning using the model as a surrogate, where the model’s predictions
are used as labels for transfer learning on audio classification models. In particular, we study a widely
distributed birdcall-specific convolutional neural network and a self-supervised neural audio codec for
encoding and decoding.
5.1. Transfer Learning
The Google Bird Vocalization Classification model is the main surrogate transfer learning experiment,
focused on version 4 of google/bird-vocalization-classifier on Kaggle, corresponding to
version 1.3 on the TensorFlow hub. We directly compute a prediction for the competition by selecting
the competition subset, filling in the missing species with negative infinity, and computing the sigmoid
of the logits. We perform fine-tuning of the model by training a new classification head on the training
dataset using the thresholded predictions of the model as pseudo-labels for the multi-label classification
task. We take advantage of the species label of the folder according to Section 5.3 as one form of
augmentation. We also fine-tune the model on the unlabeled soundscape data.
Figure 4: Diagram of the transfer learning pipeline used for experiments with BirdNET and Encodec with
the Google Bird Vocalization model as a surrogate. The soft predictions from the Bird Vocalization model are
used as pseudo-labels to train a multi-label classifier on BirdNET’s embedding space. BirdNET is also replaced
with EnCodec for comparison.
We experiment with different losses to optimize the multi-label classifier, including binary cross-
entropy (BCE), asymmetric loss (ASL), and sigmoidF1 which are explained in Section 5.4. We experi-
mented with a hidden layer to increase the capacity of the models. The model is trained for 20 epochs
with a batch size of 1,000 and a learning rate calculated by PyTorch Lightning. The training dataset
is split with an 80-20 train-validation split. The model is trained on a single NVIDIA L4 on a Google
Cloud Platform (GCP) g2-standard-8 instance with 8 vCPU, 16GB of memory, and 375GB of local NVME
storage for dataset caching and model checkpoints.
We use BirdNET V2.4 through joeweiss/birdnetlib and EnCodec through
facebookresearch/encodec v0.1.1 as comparisons for knowledge transfer. Though both
the Bird Vocalization model and BirdNET provide predictions for classification, the former provides a
more extensive set of species that overlaps with this year’s competition. We ignore the outputs of the
BirdNET model for our experiments and focus on learning the distribution of the Bird Vocalization
model’s outputs.
5.2. Data Preprocessing
We pre-compute the embeddings and predictions of the Bird Vocalization model on the training and
unlabeled soundscape datasets into a binary, columnar format that is easily accessible from network
storage. The embeddings are in a ℛ1280 space, while predictions are limited to the competition’s species
set. If the species is not present, its prediction is set to zero by assigning negative infinity to the logit
output. To save computation, we also pre-compute and join the embeddings from BirdNET and EnCodec
with the predictions of the Bird Vocalization model.
For BirdNET, we must align the model’s input size of 48kHz of 3 seconds to the 32kHz of 5 seconds
that both the Bird Vocalization model and BirdCLEF competition expect. We take the mean of the
embeddings of the 5-second audio clip with a 1-second stride for the 0th and 2nd seconds. This provides
coverage of the entire audio clip while limiting the computational burden of encoding. We take 5-second
embedding tokens at 24kHz, and limit the bandwidth of Encodec to 1.5kbps for an embedding space of
ℛ5×150 . Increasing the bandwidth to 3kpbs leads to an embedding space of ℛ5×300 . We qualitatively
inspect the embeddings through a cluster analysis in Figure 1, noting the relative difficulty of separating
common classes within the dataset.
5.3. Pseudo Multi-Label Construction
The training dataset lacks traditional labels for supervised learning, as the 5-second intervals in each
recording are not labeled with the species present. We use pseudo-labels derived from the thresholded
predictions of a surrogate model, which are not human-verified ground truth. Additionally, we use the
folder species as an extra label for further training the model. The thresholded predictions are defined
as a function of the model’s output 𝑦^ and a threshold 𝑝threshold , with the sigmoid function denoted by 𝜎.
𝑦^ = 𝜎(g(𝑥)) > 𝑝threshold (1)
We define an indicator variable 1call that determines whether the model output detects a birdcall,
which occurs when any species prediction is positive.
|𝑥|
1call (𝑥) =
∑︁
𝑥𝑖 > 0 (2)
𝑖=0
We also generate a one-hot encoding of the folder that the current audio belongs to 1species , where 𝑆
is the set of species in the folder.
{︃
1 when 𝑥 ∈ 𝑆
1species (𝑥) = (3)
0 when 𝑥 ∈/𝑆
Finally, we can define our modified label as the intersection of the model’s output and the species of
the folder. This can be implemented as a vectorized operation in PyTorch.
𝑦^species = 𝑦^ ∨ (1call (𝑦^) ∧ 1species (𝑠)) (4)
We use a threshold of 𝑝threshold = 0.5 when defining all labels. Experiments on the unlabeled
soundscapes do not have the additional information provided in the training dataset, and thus we are
limited to the pseudo-labels 𝑦^.
5.4. Training Losses
We experiment with different losses to optimize the multi-label classifier. The competition evaluation
uses a modified ROC-AUC that skips classes with no true-positive labels. We utilize MultiLabelAUROC
from the torchmetrics library as the primary learning metric. We also consider the macro-F1 score
as a secondary metric, which was utilized in the 2022 edition of the BirdCLEF competition [3]. This
metric allows us to inspect other aspects of the loss functions we consider in our experiments.
5.4.1. Binary Cross-Entropy
Binary cross-entropy is a loss function used for binary classification. It is suitable for multi-label
classification as it treats each label as an independent binary classification problem. We use this loss as
a baseline due to its simple interpretation and absence of hyperparameters.
𝑀
∑︁
𝐿=− 𝑦𝑜,𝑐 log(𝑝𝑜,𝑐 ) (5)
𝑐=1
5.4.2. Asymmetric Loss (ASL)
The asymmetric loss [13] penalizes false positives and false negatives differently. This construction
dynamically down-weights easy negative samples, hard thresholds them, and ignores misclassified
samples. This loss is well-suited for our problem domain since we have fuzzy labels from another model
initially intended for single-label classification.
𝐿+ = (1 − 𝑝)𝛾+ log(𝑝)
{︂
𝐴𝑆𝐿 = (6)
𝐿− = (𝑝𝑚 )𝛾− log (1 − 𝑝𝑚 )
The loss is defined in terms of the probability of the network output 𝑝 and hyper-parameters 𝛾+
and 𝛾− . Setting 𝛾+ > 𝛾− emphasizes positive examples while setting both terms to 0 yields binary
cross entropy. We sweep over parameters 𝛾+ ∈ {0, 1} and 𝛾− ∈ {0, 2, 4}, while the default values are
𝛾+ = 1 and 𝛾− = 4,
5.4.3. sigmoidF1
The sigmoidF1 loss [14] optimizes the F1 score directly by creating a differentiable approximation of
the F1 score. Though the competition does not score with F1, it provides a useful point of comparison
with other losses. We first define the true positive, false positive, false negative, and true negative terms
as a function of the sigmoid function.
^ ) ⊙ (1 − y)
∑︁ ∑︁
𝑡𝑝
̃︀ = S(y^ ) ⊙ y 𝑓˜𝑝 = S(y
(7)
(1 − S(y (1 − S(y ^ )) ⊙ (1 − y)
∑︁ ∑︁
𝑓˜𝑛 = ^ )) ⊙ y 𝑡𝑛 ˜ =
where S(y
^ ) is the sigmoid function applied to the model’s output y
^.
1
𝑆(𝑢; 𝛽, 𝜂) = (8)
1 + exp(−𝛽(𝑢 + 𝜂))
Then we define the F1 score as a function of the true positive, false positive, and false negative terms.
2𝑡𝑝
where (9)
̃︀
ℒ𝐹̃︁1 = 1 − 𝐹
̃︁1, 𝐹
̃︁1 =
2𝑡𝑝
̃︀ + 𝑓̃︁𝑛 + 𝑓̃︁𝑝
We are given two hyper-parameter 𝑆 = −𝛽 and 𝐸 = 𝜂. We sweep over parameters 𝑆 ∈
{−1, −15, −30} and 𝐸 ∈ {0, 1} as suggested in the author’s experiments.
6. Results
We obtain results for various models on the leaderboard via code submission on Kaggle. We report
the best validation F1 and AUROC scores, together with private and public leaderboard scores. All
submissions were made past the competition deadline with the exception of the starter Keras notebook.
We submit a model that predicts 0 for every species on the leaderboard, leading to a private and public
score of 0.5. We submit the predictions from the Bird Vocalization model and obtain a private and public
score of 0.516625 and 0.556097 respectively.
6.1. Loss Comparisons
In Table 3, we train a linear classifier head against combinations of BCE and ASL with the addition
of the species label logic. We report our validation F1 and AUROC scores alongside the private and
public scores. We note that AUROC quickly saturates against the validation set used in the training
dataset. The validation F1 score however correlates more strongly with the leaderboard scores. Using
the species labels typically increases the score by 0.05 e.g. ASL with default parameters goes from 0.529
to 0.576 in the public leaderboard.
We experimented with adding a hidden layer behind the classification head to encourage the model
to learn more complex patterns. Using ASL as the loss function, we varied the hyperparameters listed
in Table 4. We confirmed the efficacy of the species logic but noted that the scores were marginally
lower than those of the linear models. Additionally, we found that the default parameters of ASL are
effective in most tasks, with minimal tuning needed for good performance on domain-specific tasks.
Table 3
Overview of linear classifier heads on Bird Vocalization embeddings using the train dataset.
Loss Description Val F1 Val AU- Private Public
ROC Score Score
BCE labels 𝑦^ 0.5420 0.999 0.508008 0.538838
BCE labels 𝑦^𝑠𝑝𝑒𝑐𝑖𝑒𝑠 0.6095 0.997 0.546232 0.583395
ASL labels 𝑦^ 0.6035 0.999 0.523207 0.529498
ASL labels 𝑦^𝑠𝑝𝑒𝑐𝑖𝑒𝑠 0.6603 0.998 0.556189 0.576463
Table 4
An overview hyper-parameter tuning ASL on a ℛ256 hidden-layer model. Parameters of 𝛾− = 2 and 𝛾+ = 1
tend to work the best.
Description Hyperparameters Val F1 Val AU- Private Public
ROC Score Score
labels 𝑦^ 𝛾− = −2, 𝛾+ = 1 0.5727 0.996 0.557257 0.538430
labels 𝑦^ 𝛾− = −4, 𝛾+ = 0 0.5731 0.996 0.521133 0.520745
labels 𝑦^ 𝛾− = −4, 𝛾+ = 1 0.5758 0.996 0.545155 0.524194
labels 𝑦^𝑠𝑝𝑒𝑐𝑖𝑒𝑠 𝛾− = −2, 𝛾+ = 1 0.6563 0.997 0.585699 0.556193
labels 𝑦^𝑠𝑝𝑒𝑐𝑖𝑒𝑠 𝛾− = −4, 𝛾+ = 0 0.6414 0.997 0.558255 0.534317
labels 𝑦^𝑠𝑝𝑒𝑐𝑖𝑒𝑠 𝛾− = −4, 𝛾+ = 1 0.6495 0.997 0.542350 0.558353
6.2. Embedding Model Comparisons
We summarize the performance of each loss function across the CNN-based models in Table 5. Due
to CPU-time limitations on notebook runtime, we do not include an EnCodec-based model. Our best
model on the public leaderboard uses BirdNET embeddings and the BCE loss. BirdNET embeddings
consistently perform better with linear models, despite the origin of the labels being the Bird Vocalization
model. Access to the species label from the parent folder consistently improves scores. While BCE
performs well, this behavior is not indicated by our validation and private test metrics alone.
Table 5
A comparative overview of Bird Vocalization and BirdNET linear models with different losses and labeling
logic on the training dataset.
No Species With Species
Model
Private Public Private Public
Bird Vocalization (BCE) 0.508008 0.538838 0.546232 0.583395
Bird Vocalization (ASL) 0.523207 0.529498 0.556189 0.576463
Bird Vocalization (sigmoidF1) 0.529473 0.553349 0.566378 0.596767
BirdNET (BCE) 0.526441 0.577005 0.562368 0.630415
BirdNET (ASL) 0.505578 0.538719 0.550123 0.599697
BirdNET (sigmoidF1) 0.543268 0.576497 0.559693 0.590848
6.3. Dataset Comparisons
In Table 6, we compare the performance of linear models trained on the soundscape dataset using
ASL as the main loss. We observe two main results: (1) BirdNET embeddings outperform the bird
vocalization model by 0.03 on the public leaderboard and (2), models trained on the soundscape dataset
are less effective than those trained on the distribution of the training dataset. This may be attributed
to ASL’s dynamic downscaling of easily classified negative labels, making the contribution of training
labels more significant than the similarity to the test distribution.
Table 6
Scores for transfer learning on soundscapes using ASL.
Description Val F1 Val AU- Private Public
ROC Score Score
Bird Vocalization, Linear, ASL (𝛾− = 2, 𝛾+ = 1) 0.1423 0.999 0.448845 0.499735
BirdNET, Linear, ASL (𝛾− = 2, 𝛾+ = 1) 0.1158 0.997 0.47927 0.532035
Table 7
Profiling information for inference of various models, using the Python profiler via Lightning. Timing infor-
mation is collected from the predict_next step on the first 20 soundscapes sorted by identifier. The rate of
inference time per soundscape allows extrapolation to the hidden test set of 1100 4-minute recordings.
Name Profile (sec) Rate (sec/4m) Test Estimate (hours)
torchaudio 1.1 0.05 0.02
vocalization passthrough noncompiled 188.6 9.43 2.88
vocalization passthrough compiled 24.0 1.20 0.37
vocalization linear compiled 64.6 3.23 0.99
birdnet passthrough compiled 56.9 2.85 0.87
encodec passthrough noncompiled 156.4 7.82 2.39
encodec passthrough compiled 213.7 10.69 3.27
6.4. Inference Runtime
We profile each model to estimate the time required to process all test soundscapes, as shown in Table 7.
The Python profiler measures the time spent in each function and the number of function calls. Reading
all audio into chunked arrays from disk into memory, our baseline takes approximately one minute.
The Bird Vocalization model did not complete within the Kaggle’s time constraints, taking nearly three
hours according to our estimates. We compile the model using TensorFlow Lite at runtime, optimizing
operations for the hardware while allowing fallback to non-lite operations. This compilation process
results in an order-of-magnitude performance increase, leaving a substantial margin for additional
computation. The linear classification head adds only an extra half-hour of computation. The BirdNET
model also runs well within time constraints as it is compiled with TensorFlow Lite.
EnCodec exceeds the time budget, taking 2.4 hours for the base model. Experimenting with OpenVINO
[15] and applying data-independent quantization and compression did not improve inference speed.
7. Discussion
7.1. Transfer Learning Experimentation
Our transfer learning experiments using the Bird Vocalization classifier exhibit different behaviors
between the private and public leaderboards. While fine-tuned models outperform the base model when
trained on the subset of species provided for the competition, we hypothesize a shift in the species
distribution between the private and public test sets. The Bird Vocalization model is trained on a more
balanced dataset drawn from a larger set of species, whereas our transfer learning techniques rely on
pseudo-labeling from the donor model, which may not be well-calibrated for this task. We did not
account for the skew in the training data, apparent from the distribution of audio of each species.
We address label skew through different loss function choices. We use a secondary metric during
training to provide another axis to compare models. When fine-tuning the Bird Vocalization classifier
to learn the outputs from the original classifier head, the AUROC loss converges close to unity across
various architectures. However, different losses exhibit varying learning behaviors against the F1-
score, with some designed to be better surrogates than binary cross-entropy loss. During transfer
learning, these losses provide a smooth, monotonic increase to the validation F1-score, indicating that
Bird Vocalization embeddings offer a "good" representation of domain-specific data for the multi-label
problem. We observe different behaviors in other embedding spaces, supported by our clustering charts.
Figure 5: A clustering analysis of the embeddings and logits extracted from the Bird Vocalization Classifier,
on the training and soundscape data. We obtain a single vector for each track by taking the max value of
the logits and the mean value of the embeddings. The resulting vectors are clustered using PaCMAP and
demonstrate distinctive topology resulting from distinct distributional semantics.
To address skew in the training dataset, the organizers provide unlabeled soundscapes representative
of the hidden-test dataset. We discuss the distributional shift between species and frequent itemsets in
Section 4. Figure 5 shows the active intervals of calls, revealing differences in data geometry. The train
datasets at the bottom have tightly clustered logits, likely representing peaks in species probability
distributions. The embeddings form a large central cluster with several outliers, probably representing
distinctive calls. Conversely, the soundscape logit space forms two major clusters, reflecting the smaller
set of species present. Thus, soundscape embeddings should closely reflect clusters of birdcalls. It
would be interesting to explore how well we can discriminate between recording sites, as location likely
correlates with species distribution and co-occurrence patterns.
We expect soundscapes to better represent the species distribution in the hidden test set. However,
our results show that the models trained in the soundscapes perform worse than those trained on the
original dataset. Although the addition of soundscapes adds an interesting dimension to the competition,
it requires more than cursory experimentation to incorporate into modeling effectively.
7.2. Self-Supervised Neural Codecs
We find that EnCodec does not transfer well with similar experiments involving the linear and two-layer
classifiers, achieving validation F1-scores below 0.1. Adding an LSTM layer to handle the sequential
nature of EnCodec embeddings did not improve the scores. A much deeper model, similar to the
EnCodec decoder [10], is likely needed to learn from the quantized embeddings, but this is not feasible
within the competition’s inference time constraints.
Additionally, EnCodec is computationally expensive and difficult to adapt to the constrained submis-
sion environment. The Python profiler identified model inference as the bottleneck, with most time
spent on EnCodec inference. OpenVINO post-training optimizations for quantizing and compressing
weights do not significantly improve inference throughput, likely due to existing optimizations in the
upstream library. A 1.5× speedup is needed to use EnCodec in our pipeline, indicating that further
optimizations are required to leverage neural codecs based on large datasets trained with attention and
self-supervision.
8. Future Work
Figure 6: FPGrowth generates association rules from the frequent itemsets using the default minimum
confidence threshold of 0.8. Consequent itemsets are obtained with an individual species as the antecedent,
and a directed graph is created by drawing edges between items in the antecedent and consequent.
Exploiting co-occurrence species information as a prior to the learning process could be beneficial.
We have demonstrated frequent pattern mining to obtain co-occurrence distributions and quantify
differences from the training dataset. Confident relationships extracted from the data can be visualized,
as shown in Figure 6, and used to reshape the probability distribution of an existing classifier to better
represent the posterior of the unlabeled soundscape.
We aim to explore alternative parameterizations of sequential models that are computationally viable
for future competitions. The competition’s trade-offs favor compact domain-specific models over
large neural networks, focusing on linearithmic algorithms like the Fast Fourier Transform for input
representation. Finding a pre-trained neural audio codec with fewer parameters that fit within our
computational budget and pass human perceptual tests could be viable. Alternatively, training models
from scratch using different architectures via distillation methods, compatible with the encoder-decoder
architecture used in EnCodec, could be explored. State-space models like Mamba [16] provide an
appealing alternative to attention-based methods, potentially staying within our computational budget.
9. Conclusion
Our study demonstrates the effectiveness of transfer learning in birdcall classification using embeddings
from pre-trained models like Google’s Bird Vocalization Classification Model and BirdNET. These
embeddings capture meaningful structures that are beneficial for multi-label classification, although
they do not outperform many top models in the competition. Our best-performing model, which uses
BirdNET embeddings and Bird Vocalization pseudo labels to train a linear classifier, achieved a 0.63
score on the post-competition public leaderboard. Future work will focus on optimizing computational
efficiency and exploring alternative model architectures to better handle the sequential nature of audio
data. We also plan to incorporate species co-occurrence patterns to further enhance classification
accuracy. Our code is available at github.com/dsgt-kaggle-clef/birdclef-2024.
Acknowledgments
Thank you to the Data Science at Georgia Tech (DS@GT) club for providing hardware for experiments,
and to the organizers of BirdCLEF and LifeCLEF for hosting the competition.
References
[1] S. Kahl, T. Denton, H. Klinck, V. Ramesh, V. Joshi, M. Srivathsa, A. Anand, C. Arvind, H. CP,
S. Sawant, V. V. Robin, H. Glotin, H. Goëau, W.-P. Vellinga, R. Planqué, A. Joly, Overview of
BirdCLEF 2024: Acoustic identification of under-studied bird species in the western ghats, Working
Notes of CLEF 2024 - Conference and Labs of the Evaluation Forum (2024).
[2] A. Joly, L. Picek, S. Kahl, H. Goëau, V. Espitalier, C. Botella, B. Deneu, D. Marcos, J. Estopinan,
C. Leblanc, T. Larcher, M. Šulc, M. Hrúz, M. Servajean, et al., Overview of lifeclef 2024: Challenges
on species distribution prediction and identification, in: International Conference of the Cross-
Language Evaluation Forum for European Languages, Springer, 2024.
[3] S. Kahl, A. Navine, T. Denton, H. Klinck, P. Hart, H. Glotin, H. Goëau, W.-P. Vellinga, R. Planqué,
A. Joly, Overview of birdclef 2022: Endangered bird species recognition in soundscape recordings.,
in: CLEF (Working Notes), 2022, pp. 1929–1939.
[4] L. Durak, O. Arikan, Short-time fourier transform: two fundamental properties and an optimal
implementation, IEEE Transactions on Signal Processing 51 (2003) 1231–1242.
[5] S. Kahl, C. M. Wood, M. Eibl, H. Klinck, Birdnet: A deep learning solution for avian diversity
monitoring, Ecological Informatics 61 (2021) 101236.
[6] T. Denton, S. Wisdom, J. R. Hershey, Improving bird classification with unsupervised sound
separation, in: ICASSP 2022-2022 IEEE International Conference on Acoustics, Speech and Signal
Processing (ICASSP), IEEE, 2022, pp. 636–640.
[7] Y. LeCun, Y. Bengio, G. Hinton, Deep learning, nature 521 (2015) 436–444.
[8] B. Ghani, T. Denton, S. Kahl, H. Klinck, Global birdsong embeddings enable superior transfer
learning for bioacoustic classification, Scientific Reports 13 (2023) 22876.
[9] Y. Wang, H. Huang, C. Rudin, Y. Shaposhnik, Understanding how dimension reduction tools work:
An empirical approach to deciphering t-sne, umap, trimap, and pacmap for data visualization,
Journal of Machine Learning Research 22 (2021) 1–73. URL: http://jmlr.org/papers/v22/20-1061.
html.
[10] A. Défossez, J. Copet, G. Synnaeve, Y. Adi, High fidelity neural audio compression (2022).
arXiv:2210.13438.
[11] H. Klinck, M. Demkin, S. Dane, S. Kahl, T. Denton, V. Ramesh, Birdclef 2024, 2024. URL: https:
//kaggle.com/competitions/birdclef-2024.
[12] J. Han, J. Pei, Y. Yin, Mining frequent patterns without candidate generation, ACM sigmod record
29 (2000) 1–12.
[13] T. Ridnik, E. Ben-Baruch, N. Zamir, A. Noy, I. Friedman, M. Protter, L. Zelnik-Manor, Asymmetric
loss for multi-label classification, in: Proceedings of the IEEE/CVF international conference on
computer vision, 2021, pp. 82–91.
[14] G. Bénédict, V. Koops, D. Odijk, M. de Rijke, Sigmoidf1: A smooth f1 score surrogate loss for
multilabel classification, arXiv preprint arXiv:2108.10566 (2021).
[15] Y. Gorbachev, M. Fedorov, I. Slavutin, A. Tugarev, M. Fatekhov, Y. Tarkan, Openvino deep learning
workbench: Comprehensive analysis and tuning of neural networks inference, in: Proceedings of
the IEEE/CVF International Conference on Computer Vision (ICCV) Workshops, 2019.
[16] A. Gu, T. Dao, Mamba: Linear-time sequence modeling with selective state spaces, arXiv preprint
arXiv:2312.00752 (2023).