<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>Methods in Ecology and Evolution 16 (2024) 228-246.
URL: /doi/pdf/10.1111/2041-210X.14466https://onlinelibrary.wiley.com/doi/abs/10.1111/
2041-210X.14466https://besjournals.onlinelibrary.wiley.com/doi/10.1111/2041</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <article-id pub-id-type="doi">10.1111/2041-210X.14486</article-id>
      <title-group>
        <article-title>Tiles-wise Inference with Vision Transformers for Multispecies Identification in Vegetation Images⋆</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Andrea Menco-Tovar</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Jairo E. Serrano</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Juan Carlos Martinez-Santos</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Edwin Puertas</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Universidad Tecnologica de Bolivar</institution>
          ,
          <addr-line>Ternera km 1, Cartagena, Bolivar</addr-line>
          ,
          <country country="CO">Colombia</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <volume>16</volume>
      <fpage>0000</fpage>
      <lpage>0002</lpage>
      <abstract>
        <p>This paper presents a method for classifying vegetation plots containing multiple species in the context of the PlantCLEF 2025 challenge. It addresses the simultaneous identification of various species in high-resolution images using a segment-based inference approach with the Vision Transformer (ViT) model, previously pre-trained using the self-supervised learning technique DINO V2. The photos were systematically divided into diferent patch configurations, to enable accurate classification. The results showed that the optimal configuration was the 4 ×2 patch, achieving a public average macro F1 score of 0.29096 and a private score of 0.28324, ranking 13th in the challenge. Errors are observed in cases involving visually similar species, unbalanced lighting conditions, and partial species presence within the evaluated tiles. Despite these limitations, the proposed methodology confirms the potential of the ViT model in complex ecological classification tasks, highlighting the importance of future developments to improve accuracy in highly complex contexts.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Segmentation</kwd>
        <kwd>multi-species identification</kwd>
        <kwd>Vision Transformer</kwd>
        <kwd>Ecological studies</kwd>
        <kwd>Vegetation classification</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Ecological monitoring systems based on biodiversity inventories are a fundamental tool for evaluating
and managing ecosystems, enabling standardized sampling, long-term monitoring, and large-scale
analysis. Vegetation plot inventories are essential for ecological studies, as they allow for standardized
sampling, biodiversity assessment, long-term monitoring, and large-scale remote studies [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. These
vegetation inventories provide key data for biological conservation and evidence-based environmental
decision-making. Typically, inventories consist of multiple quadrat, each approximately 0.25 square
meters in size, where botanists perform a thorough visual analysis by meticulously identifying all
present species. However, these methods present notable limitations due to their high time cost and the
need for specialized expertise, which restricts the frequency and coverage of ecological studies [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], [4].
      </p>
      <p>The integration of Artificial Intelligence (AI) could significantly enhance the eficiency of specialists,
expanding the scope and reach of ecological studies. In this context, the use of AI emerges as a promising
solution to significantly improve the eficiency of environmental monitoring. It can also play a crucial
role in protecting and conserving our environment. Additionally, it provides innovative solutions
for biodiversity information systems, facilitating the online publication of data and information and
supporting comprehensive biodiversity management in a timely and eficient manner [ 5]. Today,
systems such as Pl@ntNet and iNaturalist play a crucial role in enabling users worldwide to generate,
submit, and annotate botanical observations. They also assist scientists and resource managers in
understanding where and when organisms occur. Collectively, these platforms have demonstrated
the ability to identify individual species through isolated photographs taken by citizens and scientists,
facilitating non-expert participation and large-scale data collection [6], [7].</p>
      <p>Nonetheless, the simultaneous identification of multiple species within a single high-resolution image
of vegetation plots (test images from the challenge) remains a significant technological challenge. Deep
learning models applied to plant identification require large annotated datasets. The main challenge
lies in the considerable disparity between available datasets. While training images typically represent
single-labeled individual plant images, ecological plot images encompass diverse floristic contexts, such
as those found in Pyrenean and Mediterranean floras. This challenge requires the development of robust
models that can perform efective multi-label classifications in complex ecological settings.</p>
      <p>In the PlantCLEF 2025 challenge, the evaluation focuses on multi-label prediction of plant species
in high-resolution quadrat images, where multiple species may appear, though rarely dozens
simultaneously. To measure participants’ performance, used the average per-sample macro F1 score, which
balances recall and precision, avoiding both overprediction (low precision) and underprediction (low
recall). This F1 variant first computes the score for each image independently. Then, it averages the
scores across all test set transects, sampling areas of approximately 10 m × 0.5 m at various sites, to
mitigate bias from oversampled regions and ensure a fair comparison between approaches.</p>
      <p>In general, vegetation classification requires careful examination and often involves a lengthy process.
The PlantCLEF 2025 challenge aims to predict all plant species in high-resolution plot images, making it
a multi-label classification task. This paper describes our approach and results based on our submissions
to PlantCLEF 2025. Employed the pre-trained ViT provided for this task. Adopted patch-based inference
on the test set using pre-trained models, dividing test images into non-overlapping tiles. The patch
sizes experimented with were 1×1, 4×2, 3×3, 16×16, and 14×7. Our best submission achieved a public
average per-plot macro F1 score of 0.29096 and a private average per-plot macro F1 score of 0.28324,
ranking us 13th in the challenge.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>TThis section presents relevant contributions that have employed diferent architectures and strategies
for species conservation and identification, demonstrating the potential of these methods in diverse
application contexts.</p>
      <p>It was found that, Yang et al. [8] studied the structure and dynamics of biodiversity in forest ecosystems
using forest soundscape data. They applied a deep learning-based multi-label classification approach to
ifeld recordings, successfully automating the classification of sound sources into bioacoustics, geophony,
and anthrophony. For this purpose, they designed a Convolutional Neural Network (CNN) consisting
of a convolutional feature extraction module and a fully connected classification module, achieving
macro F1 scores around 0.8421. In turn, Brun et al. [9] jointly mapped and modeled the distributions
of 2,477 plant species using Deep Neural Networks (DNNs) to assess changes in species distributions,
phenology, and dominance. They trained diferent versions of multi-species DNNs and emphasized that
multi-species DNNs predict species distributions and, especially, community composition with higher
accuracy, reporting AUC values around 0.954.</p>
      <p>In the same line, Hu et al. [10] employed four deep learning models, including a multilayer perceptron
(MLP), a CNN, a Vision Transformer (ViT), and a multimodal model to predict species and entire
community distributions, aiming to provide quantitative data for conservation eforts. They tested
the models on multispecies plant community images, obtaining macro-TSS metrics around 69.61% for
the MLP and 71.35% for the multimodal model, the latter showing better inference and fewer false
positives. Subsequently, Ghasemkhani et al. [11] presented Multi-label Federated Learning (FMLL) used
to understand and manage animal populations for biodiversity conservation and ecological management.
The proposed FMLL adopts a binary relevance strategy to handle the multi-label nature of the data and
employs the reduced error pruning tree as a classifier, achieving accuracy rates ranging from 73.24% to
94.50% on animal datasets.</p>
      <p>Continuing with the use of ViT models, the study by Elharrouss et al. [12] ofers a
comprehensive review of these models, emphasizing their fundamental principles, including the self-attention
mechanism and multi-head attention. It also highlights the versatility of these models in applications
such as image classification, medical imaging, object detection, and visual question answering while
acknowledging the challenges associated with their use, including high computational demands,
extensive data requirements, and generalization dificulties. Additionally, Saha and Xu [ 13] highlight
and explore understanding methods to optimize ViT models while also addressing their limitations.
This review highlights the application of ViT in various tasks, including image classification, object
detection, and segmentation. On the other hand, it highlights that despite their strong performance
in terms of accuracy, these models have high computational costs, memory consumption, and energy
usage.</p>
      <p>On the other hand, Lefort et al. [6] described the Pl@ntNet system, which facilitates global data
collection by allowing users to upload and annotate plant observations. They also noted that their
proposed label aggregation method aims to train AI models collaboratively for plant identification,
emphasizing the significant support that computer vision models provide in species recognition in
the field. Similarly, Van et al. [ 7] introduced a species classification and detection dataset called
iNaturalist and conducted experiments using classification and detection computer vision models,
obtaining accuracy results of 67%, which are considerably good. They concluded that state-of-the-art
computer vision models still have room for improvement when applied to large and imbalanced datasets.</p>
      <p>Overall, the reviewed works demonstrate that the use of deep architecture-based models and well
structured datasets, can significantly improve performance in species classification tasks. However,
challenges, such as addressing multi-label classification scenarios, remain.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Methodology</title>
      <p>This section details the training set used, which consists of individual observations organized by species,
as well as the test set comprising multi-species vegetation plot images captured in various floristic
contexts. Regarding the architecture, described the process of image partition, data preparation for
analysis, and the mechanism used to obtain and filter the final predictions.
3.1. Dataset Description
This section describes the characteristics and composition of the data used for both training and
evaluation. Additionally, it highlights the diferences between both sets, as well as the challenges
associated with variability in capture conditions and the complexity of the evaluated scenes.</p>
      <sec id="sec-3-1">
        <title>3.1.1. Training Set</title>
        <p>The training dataset consists of observations of individual plants, as shown in the Figure 1. These are a
subset of the Pl@ntNet training data, focusing on southwestern Europe, and cover approximately 7,800
plant species. The challenge organizers state that some reliable labels for underrepresented species
are completed using data from the GBIF platform. The images have relatively high resolution, with
a minimum or maximum of 800 pixels on the longest side, which allows for the use of classification
models capable of handling relatively high-resolution inputs and can reduce the dificulty of predicting
small plants in large vegetative plot images. The images are pre-organized into subfolders by species to
facilitate the training of individual plant classification models.</p>
        <p>There is also a complementary set of images, composed of high-resolution pseudo-square photos, as
shown in the Figure2, without labels, made available to participants to improve model adaptation to
quadrat images of multi-species vegetation.
3.1.2. Test Set
The test set comprises various datasets of plot images in diferent floristic contexts, including Pyrenean
and Mediterranean flora, as illustrated in Figure 3. Experts created these datasets, which comprise a total
of 2,105 high-resolution images. The capture protocol may vary significantly depending on the context,
including the use of wooden frames or measuring tape to delimit the plot, as well as the viewing angles
relative to the ground. Moreover, image quality may vary depending on weather conditions, which can
result in more or less pronounced shadows, blurred areas, and other efects. The main challenge of the
task lies in the domain shift between the training and test datasets. Unlike the training images, which
consist of observations of individual plants, the test images include multiple species within a single
image. Additionally, there are images containing withered plants or plants at diferent stages of growth,
as well as images with rocks, sand, and moss.</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.1.3. Architecture</title>
        <p>We worked with the pre-trained model vit_base_patch14_reg4_dinov2.lvd142m provided by
the competition, which they based on a ViT architecture pre-trained using the self-supervised learning
approach DINO V2 with 142 million images. Initially, photos from the test set, composed of
highresolution vegetation plot images, were loaded. Subsequently, files such as class_mapping, which
maps the model’s output identifiers to species identifiers, and load_species_mapping, which maps
species identifiers to recognized scientific names, were loaded. Figure 4 illustrates the entire workflow.</p>
        <p>In our approach, each image underwent a systematic spatial division process. A uniform partition
was applied to divide each original image into eight parts, arranged in four rows by two columns.
Performed this partition to obtain a more detailed and precise representation of the relevant visual
features present in each image, aiming to leverage the approach the authors used to train the ViT
model. Individually processed each resulting partition through a specific transformation that prepared
the images for further analysis. These transformations ensured that each portion met the appropriate
dimensional and normalization requirements for the ViT model.</p>
        <p>After being transformed, evaluated the images using the selected ViT model, which generated
predictions in the form of probabilities associated with various possible species for each segment. From
these probabilities, selected only those with the highest values, and performed a statistical aggregation
to determine the relative relevance of each predicted species within the total set of evaluated tiles.
Finally, it was established as a criterion that only those species whose average probabilities exceeded
5% would be considered valid for the final classification. This average probability threshold of 5% was
deliberately selected as a low value to avoid discarding species that, although appearing with low
confidence in the predictions, could actually be present in the images. This approach aims to maximize
the model’s sensitivity in the multi-species context, where partial presence or limited visibility of some
plants may generate low but relevant probabilities. These species were subsequently ranked according
to the confidence obtained.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Experimental Results</title>
      <p>The evaluation metric used by the organizers was the F1 score, designed to balance recall and precision,
ensuring that models neither overpredict nor underpredict species. Used the average per-sample macro
F1 score. Divided the results into public and private scores, and based the final rankings on the private
split. Table 1 shows the oficial public and private results of our submission. Report the average
per-sample macro F1 score and the overall ranking of our system, as well as the top-performing team
for comparison.</p>
      <p>Our system achieved a public average per-sample macro F1 score of 0.29096 and a private score of
0.28324, ranking 13th out of 45 participating teams. Compared these results with the baseline system,
which led the competition with a public average per-sample macro F1 score of 0.35900, ranking 3rd out
of 45 teams, and a private score of 0.36479, ranking 1st out of 45 teams.</p>
      <p>An increase in performance was observed when using the same model with varying numbers of
tiles 4×2, 3×3, and 14×7 as shown in Table 2. This behavior can be attributed to the fact that the 4×2
patch division allows capturing suficient distinct regions of the image while maintaining an adequate
resolution in each sub-image processed by the model. In contrast, the 14×7 sampling could introduce
redundancy or a loss of useful resolution in smaller tiles, and the 3×3 division might be insuficient to
eficiently represent areas with higher species density. It is important to highlight that the images were
divided into tiles through a uniform partition in rows and columns, resulting in rectangular tiles whose
pixel dimensions depend on the original size of each image and the chosen division configuration.</p>
      <p>No additional cropping or padding was applied in order to preserve the spatial proportion and
content of each patch. The tile configuration was primarily illustrated using the 4 ×2 format, chosen for
ofering the best performance in the conducted experiments. Although this configuration produces
rectangular tiles, the spatial integrity of the images was maintained by avoiding cropping, padding, or
stretching to force square tiles. While the main comparison reported results with non-overlapping patch
configurations, overlapping patch configurations were also explored. However, these configurations
did not improve the model’s performance in preliminary experiments and, in some cases, introduced
redundancy that negatively afected generalization.</p>
      <p>During inference on the test set, the average time per image was approximately 1.67 seconds, measured
in a computing environment with a GPU on the Google Compute Engine backend using Python 3.
Throughout this process, resource usage remained stable, with an average system RAM consumption of
4.4 GB out of 12.7 GB available, 3.6 GB of GPU memory used from 15 GB available, and a steady disk
usage of approximately 43.5 GB out of a total 112.6 GB. This indicates that the employed methodology
is feasible for applications with moderate demands in terms of processing time and computational
resource consumption. However, it is noteworthy that partitioning images into multiple tiles increases
computational cost, making it relevant to explore optimization techniques in future work to improve
eficiency without sacrificing accuracy.</p>
      <p>Heatmaps</p>
      <p>The heatmaps Figure 5 reveal the model’s maximum confidence in each patch of the evaluated
images, allowing spatial identification of areas where the model shows higher or lower certainty in its
predictions. It is observed that regions with high confidence, represented by warm colors, generally
coincide with areas where species are clearly visible and less occluded, while low confidence zones,
indicated by cool colors, correspond to sectors with shadows, dense vegetation, or visual noise factors
that hinder precise detection. These findings open the possibility of focusing future improvements
on problematic areas through specific techniques aimed at increasing the model’s robustness under
adverse conditions.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusions</title>
      <p>This paper presents our work on multi-species vegetation plot classification using the PlantCLEF 2025
dataset. Considering the challenging task involving high-resolution test images of vegetation plots,
implemented patch-based inference by dividing each plot into sizes of 1×1, 4×2, 3×3, 16×16, and 14×7,
reducing the task from multi-species identification to the prediction of one or a few classes. Achieved
our best submission using the 4×2 patch, reaching a public average per-sample macro F1 score of 0.29096
and a private score of 0.28324, ranking 13th.</p>
      <p>Additionally, observed that the model’s errors tend to concentrate on morphologically similar species
or those partially present within the analyzed tiles. In other words, the model tends to confuse species
that share close visual features, especially in contexts of high vegetation density where leaves overlap
or appear only partially. Also identified errors in images with unbalanced lighting conditions or slight
blurriness, which afects the quality of the extracted representations. In some cases, there was an
overestimation of dominant species from the training set, indicating a possible imbalance in the class
distribution learned by the model. Nevertheless, ViT has once again proven to be a competitive option
for plant identification in the context of PlantCLEF. However, further research is needed to address
resource limitations and fully leverage this robust architecture in vegetation-related tasks such as plot
classification.</p>
      <p>Finally, in future work, propose expanding the availability of labeled datasets with images of
vegetation plots that include diverse floristic contexts and varying capture conditions. It would enable the
training of more robust models adapted to complex multi-species scenarios. Additionally, it would be
relevant to investigate advanced techniques in semantic segmentation and object-based segmentation to
improve the accurate identification of individual species within each patch. It would also be feasible to
evaluate adaptive image patching methods based on criteria such as plant density or spatial distribution
to maximize the quality of extracted visual features and, consequently, model accuracy. Finally, it
is essential to examine the impact of various semi-supervised and collaborative labeling strategies,
which could significantly contribute to reducing the cost of generating annotated data and enhancing
prediction quality in ecological classification tasks.</p>
    </sec>
    <sec id="sec-6">
      <title>Generative AI Declaration</title>
      <p>During the preparation of this work, ChatGPT was used to review translation, grammar, and spelling.
After using this tool, the content, coherence, and cohesion were reviewed and edited as necessary, and
full responsibility for the content of the publication was ultimately assumed.</p>
    </sec>
    <sec id="sec-7">
      <title>Acknowledgments</title>
      <p>The authors express their gratitude to the Call 933 “Training in National Doctorates with a Territorial,
Ethnic and Gender Focus in the Framework of the Mission Policy — 2023” of the Ministry of Science,
Technology and Innovation (Minciencia). In addition, we thank the team of the Artificial Intelligence
Laboratory VerbaNex 1, afiliated with the UTB, for their contributions to this project.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>D. R. A. de Almeida</surname>
          </string-name>
          , L. B.
          <string-name>
            <surname>Vedovato</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Fuza</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Molin</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Cassol</surname>
            ,
            <given-names>A. F.</given-names>
          </string-name>
          <string-name>
            <surname>Resende</surname>
            ,
            <given-names>P. M.</given-names>
          </string-name>
          <string-name>
            <surname>Krainovic</surname>
            ,
            <given-names>C.</given-names>
            T. de Almeida, C.
          </string-name>
          <string-name>
            <surname>Amaral</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Haneda</surname>
            ,
            <given-names>R. W.</given-names>
          </string-name>
          <string-name>
            <surname>Albuquerque</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          <string-name>
            <surname>Gorgens</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Romanelli</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Ferreira</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Salk</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Espinoza</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Silva</surname>
            , E. Broadbent,
            <given-names>P. H. S.</given-names>
          </string-name>
          <string-name>
            <surname>Brancalion</surname>
          </string-name>
          ,
          <article-title>Remote sensing approaches to monitor tropical forest restoration: Current methods and future possibilities</article-title>
          ,
          <source>Journal of Applied Ecology</source>
          <volume>62</volume>
          (
          <year>2025</year>
          )
          <fpage>188</fpage>
          -
          <lpage>206</lpage>
          . doi:
          <volume>10</volume>
          .1111/
          <fpage>1365</fpage>
          -
          <lpage>2664</lpage>
          .
          <fpage>14830</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>D.</given-names>
            <surname>Sánchez-Fernández</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Jiménez-Jiménez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Fox</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. L. H.</given-names>
            <surname>Dennis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Lobo</surname>
          </string-name>
          ,
          <article-title>Identifying biodiversity hotspots over time: Stability, sampling bias, and conservation implications</article-title>
          ,
          <source>Global Ecology and Conservation</source>
          (
          <year>2025</year>
          )
          <article-title>e03586</article-title>
          . doi:
          <volume>10</volume>
          .1016/j.gecco.
          <year>2025</year>
          .e03586.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>L.</given-names>
            <surname>Picek</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Kahl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Goëau</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Adam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Larcher</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Leblanc</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Servajean</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Janoušková</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Matas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Čermák</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Papafitsoros</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Planqué</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.-P.</given-names>
            <surname>Vellinga</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Klinck</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Denton</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. S.</given-names>
            <surname>Cañas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Martellucci</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Vinatier</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Bonnet</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Joly</surname>
          </string-name>
          , Overview of lifeclef 2025:
          <article-title>Challenges on species presence prediction and identification, and individual animal identification</article-title>
          ,
          <source>in: International Conference of the Cross-Language Evaluation Forum for European Languages</source>
          , Springer,
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>