<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>Conference and Labs of the Evaluation Forum, September</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Overview of GeoLifeCLEF 2024: Species Composition Prediction with High Spatial Resolution at Continental Scale using Remote Sensing</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Lukas Picek</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Christophe Botella</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Maximilien Servajean</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>César Leblanc</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Rémi Palard</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Théo Larcher</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Benjamin Deneu</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Diego Marcos</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Joaquim Estopinan</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Pierre Bonnet</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Alexis Joly</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>AMAP, Univ Montpellier</institution>
          ,
          <addr-line>CIRAD, CNRS, INRAE, IRD, Montpellier</addr-line>
          ,
          <country country="FR">France</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>INRIA, LIRMM, Univ Montpellier</institution>
          ,
          <addr-line>CNRS, Montpellier</addr-line>
          ,
          <country country="FR">France</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>LIRMM, AMIS, Univ Paul Valéry Montpellier, Univ Montpellier</institution>
          ,
          <addr-line>CNRS</addr-line>
          ,
          <country country="FR">France</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2024</year>
      </pub-date>
      <volume>0</volume>
      <fpage>9</fpage>
      <lpage>12</lpage>
      <abstract>
        <p>Understanding the spatiotemporal distribution of species is a cornerstone of ecology and conservation. Pairing species observations with geographic and environmental predictors allows us to model the relationship between an environment and the species present at a given location. In light of that, we organize an annual competition, GeoLifeCLEF, which focuses on benchmarking and advancing state-of-the-art species distribution modeling using available bioclimatic and remote sensing data. The GeoLifeCLEF 2024 dataset spans across Europe and encompasses most of its flora. The species observation data comprises over 5 million Presence-Only (PO) occurrences and approximately 90 thousand Presence-Absence (PA) surveys. Those data are paired with various high-resolution rasters, including remote sensing imagery, land cover, and elevation, and are combined with coarse-resolution data such as climate, soil, and human footprint variables. In this paper, we present (i) an overview of the GeoLifeCLEF 2024 competition, (ii) a description of the provided data, (iii) an overview of approaches used by the participating teams, and (iv) the main results analysis.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;LifeCLEF</kwd>
        <kwd>biodiversity</kwd>
        <kwd>environmental data</kwd>
        <kwd>species distribution</kwd>
        <kwd>prediction</kwd>
        <kwd>evaluation</kwd>
        <kwd>benchmark</kwd>
        <kwd>methods comparison</kwd>
        <kwd>presence-only data</kwd>
        <kwd>presence-absence</kwd>
        <kwd>model performance</kwd>
        <kwd>remote sensing</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Global changes are transforming ecosystems at an alarming rate [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. Land use changes and biological
invasions are among the main drivers of the ongoing biodiversity decline, with their impacting processes
often operating at a fine spatial grain. Monitoring species composition at high spatial resolution (10–50m)
coherently at a continental scale would help characterize these impacts and facilitate the development
of more efective conservation policies. Unfortunately, such comprehensive monitoring is largely
impractical due to the vast areas involved, the resources required, and the complexity of ecosystems.
      </p>
      <p>
        Species distribution models (SDMs) can be used to fill the spatial gaps of field monitoring by relying
on the growing mass of spatial biodiversity data worldwide coupled with high-resolution remote sensing
data. Indeed, using spatiotemporal satellite data along with coarser geographic predictors to improve
SDM predictions has shown a potential to capture fine-scale patterns of species communities and
improve their prediction, notably through the use of deep learning-based SDM (deepSDMs) [
        <xref ref-type="bibr" rid="ref2 ref3 ref4 ref5">2, 3, 4, 5</xref>
        ].
      </p>
      <p>
        However, implementing SDMs at this resolution faces significant obstacles. The scarcity, imbalance,
and heterogeneity of available species observations and environmental data are major challenges.
Despite the mass of available Presence-Only (PO) records, notably contributed by global crowdsourcing
platforms (e.g., Pl@ntNet and iNaturalist), important sampling biases arise in their collection and
hamper the quality of SDM built from them [
        <xref ref-type="bibr" rid="ref6 ref7 ref8">6, 7, 8</xref>
        ].
      </p>
      <p>
        Last year’s edition of GeoLifeCLEF [
        <xref ref-type="bibr" rid="ref10 ref9">9, 10</xref>
        ] illustrated it by introducing an SDM evaluation scheme
based on standardized Presence-Absence (PA) surveys, and further highlighted the value of PA surveys
in model training. However, the spatial coverage of PA data being spatially very limited, properly
combining PA and PO data can leverage the extensive coverage of PO data while correcting its biases
[
        <xref ref-type="bibr" rid="ref11 ref12 ref9">9, 11, 12</xref>
        ]. Even with comprehensive data, modeling a diverse biological group as diverse as plants is
challenging. Europe has over 11,000 plant species, most of which are rare, leading to a strong class
imbalance in machine learning. This imbalance makes it dificult to develop accurate models, especially
for predicting rare species distributions. Satellite data is crucial for characterizing the spatio-temporal
context of plant communities at a high spatial resolution, and could potentially capture specific contexts
of rare species. However, the integration of satellite data into SDMs is relatively recent [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. Satellite
data brings specific challenges within SDM regarding their integration with coarser but complementary
geographic descriptors (e.g., climate, soil) and the integration of its spatial and temporal dimensions.
      </p>
      <p>For the 2024 edition of the GeoLifeCLEF1, we have assembled an extensive dataset at a European scale
to investigate these issues. This dataset is designed to facilitate the evaluation of multi-label prediction
of species composition at high spatial resolution, based on standardized Presence-Absence (PA) data
balanced across bio-regions of Europe. The campaign aims to address several key challenges:
1. Multi-label learning with massive single positive labels (i.e., PO data) and multi-label data.
2. Managing strong class imbalance; a lot of rare species with few samples and a large number of
samples for few categories.
3. Handling large-scale data and learning from multiple types of predictors, including multi-band
satellite images and time-series data.</p>
      <p>By focusing on these challenges, the GeoLifeCLEF 2024 aims to advance the field of species
distribution modeling. The integration of diverse data types and the development of robust models capable
of accurate, high-resolution predictions will provide valuable insights into species distribution
patterns. These advancements will support more efective biodiversity conservation and environmental
management practices.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Dataset and Evaluation Protocol</title>
      <p>
        The GeoLifeCLEF 2024 dataset contains species observation data, including Presence-Only occurrences
and Presence-Absence surveys, paired with various environmental predictors. This dataset ofers a
rich array of environmental rasters, Sentinel-2 satellite images, a 20-year climatic time series, and
satellite time-series point values. Building on the dataset from the previous edition [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ], we retained
the majority of the provided Presence-Only (PO) occurrences (5 million) and expanded the number of
the Presence-Absence (PA) survey records up to 90,000.
      </p>
      <p>
        As in the previous year, the PA data was divided into training and test sets (95/5) using a spatial block
hold-out procedure [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ], employing a spatial grid with 10× 10 km cells. This method ensures a thorough
evaluation of the models by randomly selecting test cells to maintain balance across biogeographical
regions. To allow easy use of the data, we provided all environmental predictors as pre-extracted scalar
values in separate CSV files. Additionally, the time-series data were formatted as 3D cubes (torch
tensors) for ease of use in machine-learning workflows.
      </p>
      <sec id="sec-2-1">
        <title>2.1. Species Observation Data</title>
        <p>
          The species observation data includes approximately 5 million Presence-Only (PO) occurrences and
around 90 thousand Presence-Absence (PA) survey records. PO data, commonly collected without
protocol, is widespread across Europe but prone to various sampling biases (see Figure 1). Reporters
(citizen scientists) may miss species due to seasonal visibility, misidentification, or lack of interest.
Below is a brief description of both PO and PA data.
1The GeoLifeCLEF 2024 competition took place in the CVPR–FGVC11 and LifeCLEF workshops [
          <xref ref-type="bibr" rid="ref13 ref14">13, 14</xref>
          ].
        </p>
        <p>
          Presence-Absence Surveys. Conducted by experienced botanists, PA surveys exhaustively report
plant species in small spatial plots (10–400 sq meters). Species not observed are likely absent. The data
originates from 29 datasets hosted in the European Vegetation Archive (EVA) and includes sources like
Denmark Naturdata, IGN National Forest Inventory, and Belgium INBOVEG. Despite the dataset’s size
(93,703 surveys), it covers only 5,016 species (about half of Europe’s flora) with highly imbalanced
species distribution. The training and test splits (95/5) were created using a spatial block hold-out
procedure [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ] with 10× 10 km cells, resulting in 88,987 training and 4,716 test surveys, ensuring
biogeographical balance.
        </p>
        <p>
          Presence-Only Occurrences. PO records are geolocated species observations with unknown sampling
protocols, ofering no information on the absence of other species. Sampling efort varies widely across
space, time, and species, often concentrating in populated areas and focusing on charismatic species.
Despite biases, PO data helps fill gaps in PA surveys when sampling biases are controlled in model
calibration [
          <xref ref-type="bibr" rid="ref11 ref12">11, 12</xref>
          ]. The PO data includes 5 million records of 9,709 species reported between 2017 and
2021 from 13 datasets extracted from the GBIF [
          <xref ref-type="bibr" rid="ref17">17</xref>
          ].
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Enviromental Predictors</title>
        <p>The spatialized geographic and environmental predictor data are crucial for precise predictive modeling.
Therefore, for each species observation (PO and PA), we provide: (i) a four-band 128× 128 satellite image
at 10 m resolution around the occurrence location, (ii) time series of past values for six satellite bands at
the point location, (iii) various environmental rasters at the European scale (e.g., climate, soil, elevation,
land use, and human footprint variables), and (iv) monthly time series of four climatic variables from
2000 to 2019. Given the dataset’s diverse sources and significant preprocessing requirements, we briefly
describe the acquisition process and data details below. Besides, we summarize all available predictors
in Table 1.
19 rasters of historical bioclim data (1981–2010)
9 pedological rasters
Elevation above sea level
According to IGBP classification (17 classes)
7 pressures on the environment for 1993 and 2009</p>
        <p>Venter et al.</p>
        <p>RGB and NIR patches centered on each observa- Sentinel-2
tion and taken the same year</p>
        <p>Source
CHELSA
CHELSA
Soilgrids
ASTER
MODIS</p>
        <p>Resolution
∼ 1km
∼ 1km
∼ 1km
∼ 30m
∼ 500m
∼ 1km
10m
30m
Satellite time series Time series of six quarterly satellite bands values Landsat</p>
        <p>since winter 1999
2.2.1. Environmental Rasters
Species observations are paired with various environmental rasters, including bioclimatic, soil,
elevation, land cover, and human footprint data. These rasters, provided as .TIF files in WGS84
(EPSG:4326) coordinates, cover Europe from (− 32.26, 26.63) to (35.58, 72.18).</p>
        <p>
          Land cover. We provide a medium-resolution (500m) multi-band land cover raster for Europe, extracted
from the MODIS Terra+Aqua dataset [
          <xref ref-type="bibr" rid="ref18">18</xref>
          ]. This GeoTIFF includes 13 layers of land cover classifications
and confidence levels. The data was processed and reprojected to WGS84, with each band describing
land cover class predictions or confidence levels. The IGBP (17 classes) and LCCS (43 classes) layers are
recommended for species distribution modeling.
        </p>
        <p>
          Human footprint data includes 16 low-resolution (1km) rasters summarizing human pressures, and
14 detailed rasters for seven pressures across two periods (1993–2009). These were derived from Venter
et al.’s global human footprint rasters [
          <xref ref-type="bibr" rid="ref19">19</xref>
          ], reprojected to WGS84. Variables include built environment,
population density, electrical infrastructure, cropland, pastureland, roads, railways, and navigable
waterways.
        </p>
        <p>Elevation data, crucial for modeling plant distribution, is provided as a GeoTIFF and in CSV format. It
was extracted from the ASTER Global Digital Elevation Model V3 via NASA portal.
Soilgrids. Nine low-resolution (1 km) soil rasters covering a depth of 5 to 15 cm were integrated from
SoilGrids 2.0. These rasters, derived from resampling the original 250m resolution data, include key soil
properties like pH and granulometry.</p>
      </sec>
      <sec id="sec-2-3">
        <title>2.3. Satellite Images</title>
        <p>
          We provide Sentinel-2 RGB and Near-Infrared (NIR) satellite images (128× 128) at 10m resolution (see
Figure 2), centered on observation locations and captured in the same year. Images are from
preprocessed rasters with cloud and shadow removal, available on the Ecodatacube platform. Values
are thresholded at 10,000, re-scaled to [
          <xref ref-type="bibr" rid="ref1">0,1</xref>
          ], gamma corrected with  = 2.5, re-scaled to [0,255], and
encoded as uint8.
        </p>
      </sec>
      <sec id="sec-2-4">
        <title>2.4. Satellite Time-Series</title>
        <p>We provide satellite time-series data spanning over 20 years, obtained from the Landsat ARD program
and pre-processed by EcoDataCube. Each location is linked to quarterly median values of six bands
(R, G, B, NIR, SWIR1, SWIR2) ranging from 1999 to 2020 and capturing environmental changes. Data
points are aggregated into CSV files and converted into 3D tensors [BAND, QUARTER, YEAR].</p>
      </sec>
      <sec id="sec-2-5">
        <title>2.5. Climatic Variables</title>
        <p>
          We provide Monthly and Average Climatic rasters from CHELSA [
          <xref ref-type="bibr" rid="ref20">20</xref>
          ]. Monthly rasters include four
variables (mean, min, max temperature, total precipitation) from Jan 2000 to Dec 2019 (960 rasters, 30
arcsec resolution). The rasters consist of 19 variables averaged from 1981 to 2010. Values for PO and PA
records are pre-extracted into CSV files and aggregated into 3D tensors [RASTER, YEAR, MONTH].
        </p>
      </sec>
      <sec id="sec-2-6">
        <title>2.6. Evaluation Metric</title>
        <p>
          As in the previous edition [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ], the evaluation metric used was the sample-averaged F1 score (F1). The
F1-score measures the agreement between predicted and actual species composition in a given area
and time. In ecological surveys, such as those in Protected Areas (PAs), each survey instance  has a
ground-truth set of labels  representing the plant species identified by experts within a grid. Given
this setup and a list of predicted labels ̂︀,1, ̂︀,2, . . . , ̂︀, , the micro 1-score is computed as follows:
F1 =
        </p>
        <p>1 ∑︁</p>
        <p>=1    + (   +   ) /2
⎧   = # of correctly predicted species, i.e., |̂︀ ∩ |.</p>
        <p>⎪
where ⎨</p>
        <p>= # of species predicted but not observed, i.e., |̂︀ ∖ |.
⎪⎩   = # of species not predicted but present, i.e., | ∖ ̂︀|.
(1)
(2)</p>
        <p>This formulation encapsulates the precision and recall elements crucial for assessing the accuracy of
predictive models in ecological studies.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Participants and methods</title>
      <p>
        A total of 83 participants / 51 teams participated2 in this year’s edition of the GeoLifeCLEF challenge
and submitted 1,184 entries, with an average of 23 entries per participant and a maximum of 175
entries by the top-ranked participant. In Figure 3, we report the private leaderboard performance for
all participants’ methods alongside the organizers’ baseline methods. Hereafter, we provide a short
overview of the teams’ methods, which are further elaborated in working notes [
        <xref ref-type="bibr" rid="ref21 ref22 ref23 ref24 ref25">21, 22, 23, 24, 25, 26</xref>
        ].
2All teams with less than 2 submissions and/or submitted only duplicated baselines were filtered out; approximately 50 teams.
      </p>
      <sec id="sec-3-1">
        <title>3.1. Baselines</title>
        <p>We provide a variety of weak and strong baselines for all participants to allow a good starting point,
continual performance increase, and working with diferent modalities. All the baselines are using PA
data only and were provided in the form of Jupyter Notebooks that could be run (both training and
inference) directly on Kaggle. Considering the significant extent to which this baseline’s performance
can be enhanced, we encouraged participants to experiment with various techniques, architectures,
losses, etc. Below, we briefly describe all baselines:
Naive baselines. With the extensive and numerous observational data available, one can naively
predict species presence by identifying the most common species within specific administrative
or biogeographical regions. For example, by predicting the top 25 most common species based on
Presence-Absence (PA) data, the resulting sample-averaged F1 score is 11.6%. In contrast, applying this
same method to Presence-Only (PO) data yields an F1 score of 8.1%. This discrepancy highlights a
distribution shift between the PA and PO data, reflecting diferences in species occurrence patterns
captured by each data collection method.</p>
        <p>Small Residual Convolutional Neural Networks for data cubes. Using a ResNet-18 architecture
[27] as a feature extractor, we have designed a model specifically tailored to fit the small cube Landsat
and Bioclimatic data (i.e., the input size of 19× 12× 4 for the bioclimatic time series and 21× 4× 6 for the
Landsat time series). The original ResNet-18 model was chosen for its balance between depth and
computational eficiency, making it a suitable candidate for our purposes. Our modifications aimed to
maintain the robustness of the original ResNet-18 model while optimizing it for the constraints of
GLC’s data dimensions. These adjustments involved reconfiguring the input layers to accommodate the
specific dimensions of the bioclimatic and Landsat datasets, ensuring that the model can efectively
capture the spatiotemporal patterns inherent in the data. The adapted models, optimized with Binary
Cross Entropy (BCE) loss, displayed significant performance by achieving sample-averaged F 1 scores of
0.259 if trained on the bioclimatic time series data and 0.266 for the Landsat time series data3. These
results underscore the efectiveness of our tailored model in dealing with GLC’s unique data structure,
suggesting that even with reduced complexity, the model retained strong predictive capabilities.
Furthermore, these findings emphasize the importance of customizing neural network architectures to
align with the specific characteristics of the input data. This ensures optimal performance without
unnecessary computational overhead and enhances model eficiency, setting a precedent for future
efforts to adapt established and state-of-the-art architectures for specialized datasets, such as GeoLifeCLEF.
Swin tranformer for the Sentinel-2 images. We made slight modifications to the architecture of a
Swin-v2-t [28] model to enable it to accept input from all four modalities of Sentinel-2 data, containing
RGB (Red, Green, Blue) and NIR (Near Infra Red) channels rather than the conventional three-channel
input. This adaptation was crucial to fully leverage the rich information provided by the Sentinel-2
dataset, which includes essential spectral data captured in the infrared range, often critical for tasks
involving vegetation and land cover analysis. Such a model achieved a sample-averaged F1 score
of 0.235. Although this score is lower compared to other models trained on diferent modalities, it
reflects the challenges and complexities involved in integrating and efectively utilizing the multimodal
data within the same model architecture. The reduction in F1 score clearly indicates potential areas
for further optimization. This includes fine-tuning the model’s hyperparameters, experimenting
with diferent training strategies, or incorporating additional preprocessing steps. Despite the
lower performance, this baseline provides valuable insights into the feasibility and limitations of
adapting transformer-based models like Swin-v2-t for multiband remote sensing data. It suggests that
while straightforward architectural modifications can enable the use of richer data inputs, achieving
optimal performance may require more sophisticated approaches to fully harness the potential of all data.
3Additionally, using the bioclimatic cubes and ResNet-18, ResNet-34, and ResNet-50, we attained sample-averaged F1 scores
of 0.251, 0.245, and 0.252, respectively.
multimodal model. Building upon the single-modality models described earlier, we created a
multimodal model that combines them. The model integrates the ResNet-18-based models for the bioclimatic
and Landsat time series data and the modified Swin-v2-t model for the Sentinel-2 data. A Multi-Layer
Perceptron (MLP) is used as a head, enabling the straightforward combination of features extracted
from all three models. This multimodal model achieved a notable sample-averaged F1 score of 0.316.
This significant improvement over the individual models’ performances highlights the inherent
multimodality of the task, demonstrating that combining diferent data sources can lead to more accurate
and robust predictions.</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Participants and Results</title>
        <p>webmaking (Top1): The best-performing participant developed and combined four types of algorithms:
(i) Predicting the  most frequent species per country, small rectangular regions, biogeographical
regions and their combination (reaching a maximum 1 = 0.21 with this approach only), (ii) Random
Forests using the previous geographic features along with bioclimatic variables (maximum 1 = 0.25
with the ensemble (i+ii)), (iii) XGBoost (maximum 1 = 0.37 with (i+ii+iii)) and (iv) our multimodal
baseline (maximum 1 = 0.41 with (i+ii+iii+iv)). Notably, the performance gains obtained by
ensembling were not only due to the combination of model types but also to the optimization of the
predicted number of species per survey and how he combined the diferent predicted species sets. For
instance, the ensemble of (i+ii) improved from 0.25 to 0.3 just by optimizing the weighting of models
in the ensemble prediction. Besides, by exploring species weighting schemes to reinforce species
predicted by several models or down-weight the predicted species pairs unlikely to co-occur among
PAs, he gained 0.02.</p>
        <p>
          AI2Lab team (Top2) [
          <xref ref-type="bibr" rid="ref21">21</xref>
          ]: This team started from the multimodal model provided as the baseline
by the organizers, to which they made several significant improvements: (i) addition of a fourth
modality (i.e., tabular environmental data encoded with an MLP), (ii) use of PO data samples through a
pseudo-labeling procedure, (iii) use of an improved encoder for the Sentinel-2 images (pre-trained
with self-supervised learning on an external dataset), (iv) use of an ensemble of models optimized on
diferent folds, and (v) optimization of the detection threshold. They finally got an F 1 score of 0.368 on
the private leaderboard. The most significant gains have been obtained by the use of the ensemble of
models (+0.021), the optimization of the detection threshold (+0.012), and the use of PO data samples
through pseudo-labeling (+0.008).
        </p>
        <p>
          Miss Qiu (Top3) [
          <xref ref-type="bibr" rid="ref24">24</xref>
          ]: This team initially reused the multimodal model baseline, but opted for a
diferent fusion method utilizing cross-attention instead of MLP. This adjustment resulted in a slight
improvement in performance. They incorporated several enhancements, some of which were similar to
the AI2Lab team’s approach (e.g., utilizing an ensemble of k-fold models), while also introducing unique
methods, such as (i) enriching predictions with species commonly found in neighboring PA and PO
samples, (ii) optimizing the number of returned species, and (iii) employing various data augmentation
techniques, including mixup. Their final F 1 score on the private leaderboard was 0.353.
BernIgen (Top5) [
          <xref ref-type="bibr" rid="ref23">23</xref>
          ]: This team started working primarily on a model using only tabular data based
on the XGBoost method (known to work very well on classical species distribution models). They have
previously reduced the dimensionality of the input data with a PCA (Principal Component Analysis)
and the number of output species by keeping only the most likely species (about 10%). This model alone
already delivers pretty good performance (F1 score of 0.31). They improved prediction performance
by adaptively predicting the number of species to return for each test plot using a regression model
(also based on XGBoost). This strategy led to a significant improvement in the F 1 score, gaining
an additional point. Finally, we combined this model with the multimodal model provided by the
organizers, achieving an F1 score of 0.349 on the private leaderboard.
        </p>
        <p>Lonan Syayf (Top9) [26]: This participant started with the provided multimodal baseline and modified
the architecture on the Landsat time-series, substituting the ResNet-based model in the baseline by a
3-layer MLP that takes the Landast time-series and bioclimatic variables as a single input vector. This
improves the private F1 from 0.316 to 0.323, with a small additional improvement to 0.329, by removing
all species with less than 10 observations. In addition, the number of species reported is made variable
by predicting as present those that have a score higher than a tunable threshold, allowing to further
improve the score to 0.342.</p>
        <p>
          Wei Dai (Top10) [
          <xref ref-type="bibr" rid="ref22">22</xref>
          ]: This team used the provided multimodal baseline and focused primarily on
the least performing modality – Sentinel-2 image patches. They tried diferent architectures (e.g.,
ConvNeXt [29], MaxVit [30], and Swin-v2 [28]) and their ensembling. By ensembling four models, they
reduced the relative error of the provided baseline by around 15%; the best single model – MaxVit-t –
reduced the relative error by 12.5%.
        </p>
        <p>
          DS@GT-GeoLifeCLEF (Top48) [
          <xref ref-type="bibr" rid="ref25">25</xref>
          ]: This team explored various methods, including (i) eficient nearest
neighbor search using Locality Sensitive Hashing, (ii) training convolutional neural networks on DCT
coeficients instead of raw pixels, and (iii) utilizing Tile2Vec [ 31], a self-supervised learning technique
that generates embeddings of satellite imagery tiles. They encountered and reported various dificulties
with model convergence and could not achieve results surpassing an F1 score of 0.16.
        </p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Discussion and Conclusion</title>
      <p>The main outcomes we can derive from the GeoLifeCLEF 2024 are the following:
Provided baselines had a positive impact on overall performance. Compared to last year,
with a single team outperforming the best baseline, this year, 25 participants achieved similar or
better performance than the provided baselines. We assume that the continuous process of
publishing better baselines increased the participation engagement since allowed continuous improvements.
Proactive engagement with the community and continual release of better baselines increased
the impact. Similarly, as in the previous case, the proactive engagement with the community through
the Kaggle forum allowed crowd-sourcing of methods and continuous incremental performance
increase. Compared to other LifeCLEF and FGVC competitions, the GeoLifeCLEF 2024 competition got
10–100 times more participants and submissions.</p>
      <p>Multimodal is more than single-modal. The multimodal models were the key to success. All
participants’ working notes reported large performance gains by combining models based on diferent
modalities, regardless of the type of algorithm used to exploit these modalities (e.g., random forests,
XGBoost, CNNs). Yet, the potential of the high dimensional remote sensing data was only exploited by
deep learning-based models.</p>
      <p>Accounting for species community capacity is key. Our baseline development showed that the
sum of species presence probabilities predicted independently of each other largely over-predicted the
actual number of species. Most top teams developed techniques to constrain or predict the number of
present per species community before applying a probability ranking rule [32]. Given the high spatial
resolution considered here, it is coherent with the important local constraints of species assemblages,
such as competition.</p>
      <p>High amount of Presence-Absence (PA) training data positively influence the results . Methods
based on the PA data consistently outperformed the ones based solely on the PO data. Some minor
gains were reported in combining PA and PO data through specific methods accounting for sampling
biases. The provision of more PA data in the training dataset may have contributed to a much higher
performance compared to last year’s edition (for which the best F1 score was 0.27). Many challenges
remain in developing data integration methods compatible with machine learning modeling pipelines.</p>
      <p>For the future, it seems important to understand why improving performance with Presence-Only
data is dificult, even though it is much larger. The presence of observation bias is clearly a plausible
reason (some species are observed more than others), but it seems the spatial scale of the test set’s plots
may also be an issue. They are indeed quite small (10× 10m on average) and do not necessarily reflect
the presence of all the species in larger areas such as the one considered by the models. Moreover, the
locations of these plots themselves follow specific protocols, which may introduce observation biases
diferent from those of Presence-Only data.</p>
      <p>Interestingly, the development of multimodal models highlights the importance of leveraging diverse
data inputs in environmental monitoring and analysis. This approach not only sets a promising precedent
but also opens up exciting possibilities for future work in the field, suggesting that multimodal data
fusion can substantially enhance the performance of predictive models in complex, real-world tasks. It
also highlights the inherent dificulty in balancing the contributions of each modality to the overall
prediction accuracy, especially when dealing with high-dimensional and diverse data inputs. We
recognize these challenges and are committed to addressing them in our future work.</p>
    </sec>
    <sec id="sec-5">
      <title>Acknowledgement</title>
      <p>The research described in this paper was funded by the European Commission via the MAMBO
(http:doi.org/10.3030/101060639) and GUARDEN (http:doi.org/10.3030/101060693) projects, which have
received funding from the European Union’s Horizon Europe research and innovation program under
grant agreements 101060693 and 101060639.
Evaluation Forum, 2024.
[26] A. Syayfetdinov, Multimodal networks for species distribution modeling, in: Working Notes of</p>
      <p>CLEF 2024 - Conference and Labs of the Evaluation Forum, 2024.
[27] K. He, X. Zhang, S. Ren, J. Sun, Deep residual learning for image recognition, in: Proceedings of
the IEEE conference on computer vision and pattern recognition, 2016, pp. 770–778.
[28] Z. Liu, H. Hu, Y. Lin, Z. Yao, Z. Xie, Y. Wei, J. Ning, Y. Cao, Z. Zhang, L. Dong, et al., Swin
transformer v2: Scaling up capacity and resolution, in: Proceedings of the IEEE/CVF conference
on computer vision and pattern recognition, 2022, pp. 12009–12019.
[29] Z. Liu, H. Mao, C.-Y. Wu, C. Feichtenhofer, T. Darrell, S. Xie, A convnet for the 2020s, in:
Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2022, pp.
11976–11986.
[30] Z. Tu, H. Talebi, H. Zhang, F. Yang, P. Milanfar, A. Bovik, Y. Li, Maxvit: Multi-axis vision
transformer, in: European conference on computer vision, Springer, 2022, pp. 459–479.
[31] N. Jean, S. Wang, A. Samar, G. Azzari, D. Lobell, S. Ermon, Tile2vec: Unsupervised representation
learning for spatially distributed data, in: Proceedings of the AAAI Conference on Artificial
Intelligence, volume 33, 2019, pp. 3967–3974.
[32] M. D’Amen, A. Dubuis, R. F. Fernandes, J. Pottier, L. Pellissier, A. Guisan, Using species richness and
functional traits predictions to constrain assemblage predictions from stacked species distribution
models, Journal of biogeography 42 (2015) 1255–1266.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>E. S.</given-names>
            <surname>Brondizio</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Settele</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Diaz</surname>
          </string-name>
          , H. T. Ngo,
          <article-title>Global assessment report on biodiversity and ecosystem services of the intergovernmental science-policy platform on biodiversity and ecosystem services, IPBES secretariat</article-title>
          , Bonn, Germany. (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>C.</given-names>
            <surname>Botella</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Joly</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Bonnet</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Monestiez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Munoz</surname>
          </string-name>
          ,
          <article-title>A deep learning approach to species distribution modelling, Multimedia Tools and Applications for Environmental</article-title>
          &amp; Biodiversity
          <string-name>
            <surname>Informatics</surname>
          </string-name>
          (
          <year>2018</year>
          )
          <fpage>169</fpage>
          -
          <lpage>199</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>B.</given-names>
            <surname>Deneu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Servajean</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Bonnet</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Botella</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Munoz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Joly</surname>
          </string-name>
          ,
          <article-title>Convolutional neural networks improve species distribution modelling by capturing the spatial structure of the environment</article-title>
          ,
          <source>PLoS computational biology 17</source>
          (
          <year>2021</year>
          )
          <article-title>e1008856</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>B.</given-names>
            <surname>Deneu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Joly</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Bonnet</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Servajean</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Munoz</surname>
          </string-name>
          ,
          <article-title>Very high resolution species distribution modeling based on remote sensing imagery: how to capture fine-grained and large-scale vegetation ecology with convolutional neural networks?</article-title>
          ,
          <source>Frontiers in plant science 13</source>
          (
          <year>2022</year>
          )
          <fpage>839279</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>J.</given-names>
            <surname>Estopinan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Servajean</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Bonnet</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Munoz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Joly</surname>
          </string-name>
          ,
          <article-title>Deep species distribution modeling from sentinel-2 image time-series: a global scale analysis on the orchid family</article-title>
          ,
          <source>Frontiers in Plant Science</source>
          <volume>13</volume>
          (
          <year>2022</year>
          )
          <fpage>839327</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>E. H.</given-names>
            <surname>Boakes</surname>
          </string-name>
          ,
          <string-name>
            <surname>P. J. McGowan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. A.</given-names>
            <surname>Fuller</surname>
          </string-name>
          ,
          <string-name>
            <surname>D.</surname>
            Chang-qing,
            <given-names>N. E.</given-names>
          </string-name>
          <string-name>
            <surname>Clark</surname>
          </string-name>
          ,
          <string-name>
            <surname>K. O'Connor</surname>
            ,
            <given-names>G. M.</given-names>
          </string-name>
          <string-name>
            <surname>Mace</surname>
          </string-name>
          ,
          <article-title>Distorted views of biodiversity: spatial and temporal bias in species occurrence data</article-title>
          ,
          <source>PLoS biology 8</source>
          (
          <year>2010</year>
          )
          <article-title>e1000385</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>N. J.</given-names>
            <surname>Isaac</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. J.</given-names>
            <surname>Pocock</surname>
          </string-name>
          ,
          <article-title>Bias and information in biological records</article-title>
          ,
          <source>Biological Journal of the Linnean Society</source>
          <volume>115</volume>
          (
          <year>2015</year>
          )
          <fpage>522</fpage>
          -
          <lpage>531</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>T.</given-names>
            <surname>Mesaglio</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. T.</given-names>
            <surname>Callaghan</surname>
          </string-name>
          ,
          <article-title>An overview of the history, current contributions and future outlook of inaturalist in australia</article-title>
          ,
          <source>Wildlife Research</source>
          <volume>48</volume>
          (
          <year>2021</year>
          )
          <fpage>289</fpage>
          -
          <lpage>303</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>C.</given-names>
            <surname>Botella</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Deneu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. M.</given-names>
            <surname>Gonzalez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Servajean</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Larcher</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Leblanc</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Estopinan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Bonnet</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Joly</surname>
          </string-name>
          , Overview of geolifeclef 2023:
          <article-title>Species composition prediction with high spatial resolution at continental scale using remote sensing</article-title>
          ,
          <source>in: CLEF 2023: Conference and Labs of the Evaluation Forum</source>
          ,
          <year>2023</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>C.</given-names>
            <surname>Botella</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Deneu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. Marcos</given-names>
            <surname>Gonzalez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Servajean</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Larcher</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Leblanc</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Estopinan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Bonnet</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Joly</surname>
          </string-name>
          , Overview of GeoLifeCLEF 2023:
          <article-title>Species composition prediction with high spatial resolution at continental scale using remote sensing</article-title>
          ,
          <source>in: Working Notes of CLEF 2023 - Conference and Labs of the Evaluation Forum</source>
          ,
          <year>2023</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>W.</given-names>
            <surname>Fithian</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Elith</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Hastie</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. A.</given-names>
            <surname>Keith</surname>
          </string-name>
          ,
          <article-title>Bias correction in species distribution models: pooling survey and collection data for multiple species</article-title>
          ,
          <source>Methods in Ecology and Evolution</source>
          <volume>6</volume>
          (
          <year>2015</year>
          )
          <fpage>424</fpage>
          -
          <lpage>438</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>D. A.</given-names>
            <surname>Miller</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Pacifici</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. S.</given-names>
            <surname>Sanderlin</surname>
          </string-name>
          ,
          <string-name>
            <surname>B. J. Reich,</surname>
          </string-name>
          <article-title>The recent past and promising future for data integration methods to estimate species' distributions</article-title>
          ,
          <source>Methods in Ecology and Evolution</source>
          <volume>10</volume>
          (
          <year>2019</year>
          )
          <fpage>22</fpage>
          -
          <lpage>37</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>A.</given-names>
            <surname>Joly</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Picek</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Kahl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Goëau</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Espitalier</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Botella</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Deneu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Marcos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Estopinan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Leblanc</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Larcher</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Šulc</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Hrúz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Servajean</surname>
          </string-name>
          , et al.,
          <source>Overview of LifeCLEF</source>
          <year>2024</year>
          :
          <article-title>Challenges on species distribution prediction and identification</article-title>
          ,
          <source>in: International Conference of the Cross-Language Evaluation Forum for European Languages</source>
          , Springer,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>A.</given-names>
            <surname>Joly</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Picek</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Kahl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Goëau</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Espitalier</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Botella</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Deneu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Marcos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Estopinan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Leblanc</surname>
          </string-name>
          , et al.,
          <article-title>Lifeclef 2024 teaser: Challenges on species distribution prediction and identification</article-title>
          ,
          <source>in: European Conference on Information Retrieval</source>
          , Springer,
          <year>2024</year>
          , pp.
          <fpage>19</fpage>
          -
          <lpage>27</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>C.</given-names>
            <surname>Botella</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Benjamin</surname>
          </string-name>
          , M. Diego Gonzalez,
          <string-name>
            <given-names>S.</given-names>
            <surname>Maximilien</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Théo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Joaquim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>César</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Bonnet</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Joly</surname>
          </string-name>
          ,
          <article-title>The GeoLifeCLEF 2023 Dataset to evaluate plant species distribution models at high spatial resolution across</article-title>
          <source>Europe</source>
          ,
          <year>2023</year>
          . URL: https://hal.science/hal-04152362,
          <article-title>the full dataset is freely available at the link below (perennial repository) for academic use or other non-commercial use</article-title>
          : https://lab.plantnet.org/seafile/d/936fe4298a5a4f4c8dbd/.
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>D. R.</given-names>
            <surname>Roberts</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Bahn</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Ciuti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. S.</given-names>
            <surname>Boyce</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Elith</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Guillera-Arroita</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Hauenstein</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. J.</given-names>
            <surname>LahozMonfort</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Schröder</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Thuiller</surname>
          </string-name>
          , et al.,
          <article-title>Cross-validation strategies for data with temporal, spatial, hierarchical, or phylogenetic structure</article-title>
          ,
          <source>Ecography</source>
          <volume>40</volume>
          (
          <year>2017</year>
          )
          <fpage>913</fpage>
          -
          <lpage>929</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>GBIF.Org</given-names>
            <surname>User</surname>
          </string-name>
          , Occurrence download,
          <year>2022</year>
          . URL: https://www.gbif.org/occurrence/download/ 0144742-
          <fpage>220831081235567</fpage>
          . doi:
          <volume>10</volume>
          .15468/DL.
          <year>8WVZQF</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>M. A.</given-names>
            <surname>Friedl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Sulla-Menashe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Tan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Schneider</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Ramankutty</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Sibley</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <article-title>Modis collection 5 global land cover: Algorithm refinements and characterization of new datasets</article-title>
          ,
          <source>Remote sensing of Environment</source>
          <volume>114</volume>
          (
          <year>2010</year>
          )
          <fpage>168</fpage>
          -
          <lpage>182</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>O.</given-names>
            <surname>Venter</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E. W.</given-names>
            <surname>Sanderson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Magrach</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. R.</given-names>
            <surname>Allan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Beher</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. R.</given-names>
            <surname>Jones</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H. P.</given-names>
            <surname>Possingham</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W. F.</given-names>
            <surname>Laurance</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Wood</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B. M.</given-names>
            <surname>Fekete</surname>
          </string-name>
          , et al.,
          <article-title>Global terrestrial human footprint maps for 1993 and 2009, Scientific data 3 (</article-title>
          <year>2016</year>
          )
          <fpage>1</fpage>
          -
          <lpage>10</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>D. N.</given-names>
            <surname>Karger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Conrad</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Böhner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Kawohl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Kreft</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. W.</given-names>
            <surname>Soria-Auza</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N. E.</given-names>
            <surname>Zimmermann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H. P.</given-names>
            <surname>Linder</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Kessler</surname>
          </string-name>
          ,
          <article-title>Climatologies at high resolution for the earth's land surface areas</article-title>
          ,
          <source>Scientific data 4</source>
          (
          <year>2017</year>
          )
          <fpage>1</fpage>
          -
          <lpage>20</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Peng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.-S.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <article-title>Combining present-only and present-absent data with pseudolabel generation for species distribution modeling</article-title>
          ,
          <source>in: Working Notes of CLEF 2024 - Conference and Labs of the Evaluation Forum</source>
          ,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Cheng</surname>
          </string-name>
          , W. Dai,
          <string-name>
            <surname>J.</surname>
          </string-name>
          <article-title>Sun, Multi-modal feature fusion networks for geolifeclef 2024</article-title>
          ,
          <source>in: Working Notes of CLEF 2024 - Conference and Labs of the Evaluation Forum</source>
          ,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>T.</given-names>
            <surname>Chopard</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Rawlings</surname>
          </string-name>
          ,
          <article-title>Exploring biodiversity: A multi-model approach to multi-label plant species prediction</article-title>
          ,
          <source>in: Working Notes of CLEF 2024 - Conference and Labs of the Evaluation Forum</source>
          ,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>H.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Tao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Jiang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Wan</surname>
          </string-name>
          ,
          <article-title>Tighnari: Multi-modal plant species prediction based on hierarchical cross-attention using graph-based and vision backbone-extracted features</article-title>
          ,
          <source>in: Working Notes of CLEF 2024 - Conference and Labs of the Evaluation Forum</source>
          ,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>A.</given-names>
            <surname>Miyaguchi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Aphiwetsa</surname>
          </string-name>
          ,
          <string-name>
            <surname>M.</surname>
          </string-name>
          <article-title>McDufie, Tiled compression and embeddings for multilabel classification in geolifeclef 2024</article-title>
          , in: Working Notes of CLEF 2024 - Conference and Labs of the
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>