Overview of FungiCLEF 2024: Revisiting Fungi Species Recognition Beyond 0-1 Cost Lukas Picek1,2,* , Milan Šulc3 and Jiří Matas4 1 Department of Cybernetics, Faculty of Applied Sciences, University of West Bohemia, Czech Republic 2 Inria, LIRMM, University of Montpellier, CNRS, France 2 Second Foundation, Czech Republic 3 The Center for Machine Perception Dept. of Cybernetics, FEE, Czech Technical University in Prague, Czech Republic Abstract The third edition of the fungi recognition challenge, FungiCLEF 2024, organized within LifeCLEF, advances the field of mushroom species identification using computer vision and machine learning. Building on the Danish Fungi 2020 dataset and incorporating new data from the CzechFungi app, FungiCLEF 2024 challenges participants to recognize fungi species from images and metadata, focusing on efficient inference and minimalization of edible and poisonous species confusion. The strict limits on computational complexity ensure that the resulting solutions are practical for use in real-world settings with limited computational resources. The competition attracted seven teams, with five outperforming the provided baseline, which was based on the pre-trained EfficientNet-B1 model. This overview paper provides (i) a comprehensive description of the challenge and provided baseline method, (ii) detailed characteristics of the dataset and task specifications, (iii) an examination of the methods employed by contestants, and (iv) a discussion of the competition outcomes. The results highlight incremental advancements in fungi recognition, showcasing innovative approaches and techniques that push the limits of previous work. Keywords LifeCLEF, FungiCLEF, fine-grained visual categorization, metadata, open-set recognition, fungi, species identifica- tion, machine learning, computer vision, classification 1. Introduction Fungi recognition systems based on computer vision and machine learning [1, 2] are transforming the field of mycology, making it easier than ever for researchers, enthusiasts, and professionals to identify mushroom species. Tasks that once required extensive expertise, i.e., studying the existing literature, can now be accomplished in a few seconds. The fungi identification service offered by the Atlas of Danish Fungi [3] exemplifies this advancement: users simply capture an image of their specimen, and the system promptly generates a list of probable species matches. This facilitates efficient manual verification by allowing users to compare their observations with reference photos and species descriptions. Additionally, it encourages users to contribute valuable biodiversity observations, enhancing the overall understanding and documentation of fungal diversity. Despite the impressive performance of automatic fungi species recognition systems, significant challenges remain due to the complexity and diversity of fungal species. One major challenge is the vast number of fine-grained categories (species) that exist. Many of these species exhibit high visual similarities, making it difficult to distinguish between them even though they may not be genetically related (see Figure 1). This visual similarity can easily lead to misidentification, as the algorithm might not reliably discern subtle differences in color, shape, or texture. Additionally, there is significant intra-class variance within species. The appearance of fungal specimens can vary widely based on several factors, including genotype, age, seasonal conditions, and the local environment. For instance, a mushroom of the same species can look markedly different when it is young compared to when it is mature. Seasonal variations can affect the color and size of fungi, while local environmental conditions CLEF 2024: Conference and Labs of the Evaluation Forum, September 09–12, 2024, Grenoble, France * Corresponding author. $ picekl@kky.zcu.cz (L. Picek)  0000-0002-6041-9722 (L. Picek); 0000-0002-6321-0131 (M. Šulc); 0000-0003-0863-4844 (J. Matas) © 2024 Copyright for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0). CEUR ceur-ws.org Workshop ISSN 1613-0073 Proceedings Russula Russula Russula Hortiboletus Suillellus Neoboletus Amanita Amanita Amanita emetica paludosa rosea rubellus queletii luridiformis muscaria rubescens pantherina Figure 1: Fungi species intra- and inter-class similarities. Visual similarities and differences across nine selected species and three distinct families (i.e., higher taxonomic ranks). Image taken from [8]. such as humidity, light exposure, and soil type can further influence their appearance [4, 5]. These variations pose a considerable challenge for automatic recognition systems, which must be robust enough to account for such differences to ensure accurate identification. Moreover, the quality and resolution of images submitted for recognition can vary significantly, impacting the system’s ability to accurately classify the observations. Images taken in the wild might suffer from poor lighting, occlusions, or background noise, adding another layer of complexity to the recognition task [6, 7]. These challenges underscore the importance of developing new methods capable of handling the complex details of fungal diversity, pushing the boundaries of current state-of-the-art in fungi species recognition, and fine-grained visual categorization in general. To allow continual incremental improvements in fungi recognition, we organize an annual research competition – FungiCLEF. The latest edition, which took part of the LifeCLEF 2024 [9] and the FGVC11 workshop at CVPR 2024, builds on the foundations laid by its predecessors [10, 11]. This year’s challenge introduces an updated dataset and continues to emphasize model efficiency by imposing computational and memory constraints. The competition presents participants with fungi recognition scenarios that address real-world applications, including fungi species identification and distinguishing between poisonous and edible mushrooms. By promoting innovation in this field, FungiCLEF 2024 aims to bridge further the gap between computer vision capabilities and practical mycological needs, potentially impacting areas ranging from biodiversity research to public health and beyond. 2. Challenge Description Efficient and scalable species recognition is essential for large-scale initiatives such as citizen science projects [1, 2], which often operate with limited computational resources. Species identification typically relies not just on the visual features of the specimen but also on additional contextual data about habitat, substrate, location, etc. For the FungiCLEF2024 competition, we have developed a benchmark with rich metadata and expert-verified labels for testing the performance of systems that combine visual and contextual information. Given that mushrooms are often foraged for consumption, the competition also addresses scenarios related to misclassification between edible and poisonous species. This ensures robust outcomes in fungal species recognition, enhancing scientific research and public safety. To enable use in practical applications, all participants had to submit their models via the HuggingFace evaluation platform and must pass computational limits. Each classification model had to fit limits for prediction time limit (120 minutes) within a given HuggingFace server instance (Nvidia T4 small 4vCPU, 15GB RAM, 16GB VRAM). Table 1 FungiCLEF 2024 dataset statistics. We list the number of species and images in each subset. The known and unknown represent how many species were and were not available in the training dataset. Subset # of Species → known/unknown # of Images # of Observations Training 1,604 1,604 / – 295,938 177,170 Validation 3,299 1,084 / 1,629 91,231 45,021 Test 1,398 749 / 649 41,177 22,412 CzechFungi App 137 94 / 43 393 215 Atlas of Danish Fungi 1,261 721 / 540 40,784 22,197 2.1. Dataset The FungiCLEF 2024 dataset builds upon the previous editions of FungiCLEF [12, 13], LifeCLEF [14, 15, 16], and the Danish Fungi 2020 dataset [8]. All training data is derived from a citizen science platform – the Atlas of Danish Fungi. Each fungi observation in the provided dataset has undergone expert validation, ensuring high-quality species labels. The dataset features rich observation metadata, i.e., information about habitat, substrate, timestamp, location, etc. Provided subsets (i.e., training, validation, and test) are briefly described below, and their statistics in detail are listed in Table 1. The training set is based on 177,170 real and expert-verified fungi observations of 1,604 species. The dataset is built exclusively from the Danish Fungi 2020 data by combining the training and public test sets and includes 295,938 images. The validation set comprises expert-validated observations with species labels collected solely in 2022. This subset includes 3,299 fungi species and contains 45,021 observations with many "unknown" species. The test set is based on two subsets originating from two applications (Atlas of Danish Fungi and CheckFungi) and two countries, Denmark and the Czech Republic respectively. The CheckFungi dataset is a small subset containing only around 200 submissions and is included primarily as a control set to prevent cheating. The test set was split 80/20 for public and private evaluation, respectively. 2.2. Evaluation Protocol Same as the previous year [13], the challenge aims at predicting the species given visual observation and metadata and considers scenarios focusing on correct species classification as well as on classifying edible vs poisonous mushrooms. Namely, the goal is to minimize the empirical loss 𝐿 for decisions 𝑞(𝑥) over observations 𝑥 and true labels 𝑦, given a cost function 𝑊 (𝑦, 𝑞(𝑥)). ∑︁ 𝐿= 𝑊 (𝑘𝑖 , 𝑞(𝑥𝑖 )). (1) 𝑖 Different recognition scenarios and their cost function 𝑊 (𝑦, 𝑞(𝑥)) are described together with their motivation in the points below: • Track 1: Standard classification with "unknown" category (an open-set scenario). The first metric was the standard classification accuracy, i.e., the average correctness of the predicted class. All species not represented in the training set had to be correctly classified as an "unknown" category. The decision function is simple: each observation is simply represented by an identity matrix, i.e., 0 if 𝑞(𝑥) = 𝑦 {︂ 𝑊1 (𝑦, 𝑞(𝑥))) = (2) 1 otherwise • Track 2: Cost for confusing edible species for poisonous and vice versa. Let us have a function 𝑑 that indicates dangerous (poisonous) species as 𝑑(𝑦) = 1 if species 𝑦 is poisonous, and 𝑑(𝑦) = 0 otherwise. Let us denote 𝑐PSC the cost for poisonous species confusion (if a poisonous observation was misclassified as edible) and 𝑐ESC the cost for edible species confusion (if an edible observation was misclassified as poisonous). if 𝑑(𝑦) = 𝑑(𝑞(𝑦)) ⎧ ⎨ 0 𝑊2 (𝑦, 𝑞(𝑥))) = 𝑐PSC if 𝑑(𝑦) = 1 and 𝑑(𝑞(𝑦)) = 0. (3) 𝑐ESC otherwise ⎩ For the benchmark, we set 𝑐ESC = 1 and 𝑐PSC = 100. • Track3: A user-focused loss composed of both the classification error and the poi- sonous/edible confusion. Assuming the user is interested both in the species classification as well as in low edible to poisonous species confusion (and vice versa), the third cost function simply combines 𝑊1 and 𝑊2 : 𝑊3 (𝑦, 𝑞(𝑥))) = 𝑊1 (𝑦, 𝑞(𝑥)) + 𝑊2 (𝑦, 𝑞(𝑥)). (4) 2.3. Baseline To enable an easier start for all participants and straightforward model evaluation, we provide a weak baseline based on the pre-trained PyTorch EfficientNet-B1 [17] model wrapped into a HuggingFace repository, allowing direct evaluation on the private test set. This repository includes the model weights and inference scripts. In addition to the PyTorch-based submission, we offer an example for submitting an ONNX model. This ONNX model was initially provided for another LifeCLEF [9, 18] competition, SnakeCLEF 2024 [19]. The pre-trained EfficientNet-B1 model used in our baseline was originally published in the Danish Fungi 2020 dataset [8]. This model has demonstrated relatively strong performance in fungi species classification and serves as a robust starting point. Overall, our goal was to offer a comprehensive and accessible starting point for all participants, enabling them to focus on developing novel solutions and improving upon the provided baseline. 2.4. Timeline The FungiCLEF 2024 competition was launched on March 13, 2024, and was promoted through the LifeCLEF, HuggingFace, and FGVC challenge web pages, inviting participants to register. The com- petition ran for approximately three months, with the final submission deadline on May 24. Similar to the previous year, the test data remained confidential. Participants were allowed to make up to five submissions per day using the HuggingFace evaluation platform to assess their models on the test set. Two weeks before the deadline, the submission limit was increased to ten per day. After the competition concluded, all participants had the opportunity to submit post-competition entries for further evaluation of their ablations. 2.5. Working Notes Participants were strongly encouraged to submit both their code and a detailed technical report (Working Notes) to ensure their results can be fully reproduced. All the submitted Working Notes underwent thorough review and were given complex feedback by 2-3 experts with extensive publication records in Computer Vision and Machine Learning. This rigorous review process was designed to guarantee reproducibility and maintain quality standards. The review was single-blind, allowing participants to respond with up to two rebuttals to address any feedback or concerns raised by the reviewers. These working notes provide an in-depth analysis of the techniques employed, including hyperparameter tuning, model ensembling, and loss function selection, offering valuable insights into the development of the method for fungal image classification. 3. Challenge Results This year, the three tracks of FungiCLEF have three different best-performing submissions by three different teams 1 . However, for the official ranking, the Track 3 score was selected. The best-performing submission in Track 1 by Jack Etheredge [20] achieved a score of 0.240. The best scores in Track 2 were achieved by team upupup [21] with a score of 0.072. Finally, the best score for Track 3, the main competition track, was achieved by team IES [22] with a score of 0.362. The official challenge results, in terms of Track 1, Track 2, and Track 3 metrics, are reported in Figure 2. For completeness, we note that a post-competition submission by team DS@GT [23] achieved even higher recognition scores, highlighting the competitive and evolving nature of the challenge. This post-competition effort serves as a testament to the ongoing advancements and innovations in the field, extending beyond the official competition period. 0.6 2.5 1.604 1.75 0.500 0.500 2.044 0.5 1.50 2.0 0.391 0.377 0.4 1.25 Track1 Metric Track3 Metric Track2 Metric 1.5 1.153 1.153 0.276 0.275 0.269 0.3 1.00 0.240 0.651 0.75 0.651 1.0 0.693 0.667 0.2 0.438 0.513 0.396 0.50 0.407 0.362 0.1 0.5 0.25 0.082 0.081 0.072 0.0 0.00 0.0 chirmy chirmy TingTing1999 IES glhr upupup IES TingTing1999 glhr upupup jack-etheredge DS@GT jack-etheredge DS@GT chirmy TingTing1999 glhr upupup IES jack-etheredge DS@GT EfficientNet-B1 EfficientNet-B1 EfficientNet-B1 Figure 2: Private Leaderboard of the FungiCLEF 2024 competition. We report performance for all 7 teams. The track metrics are recognition losses, thus lower is better. Each track could have a different submission; therefore, Track1 and Track2 scores for the same team do not sum up to the Track3 score. The orange color depicts baseline performance. 4. Participants and Methods This year, a total of seven teams participated in the FungiCLEF 2024 challenge; of these, five outperformed the baseline EfficientNet-B1 in Track 3, and five submitted working notes from which four passed the review process and were accepted for publication. The methodologies varied and included a range of techniques from state-of-the-art neural network architectures to sophisticated strategies to incorporate the metadata. Details of the best methods and systems used are synthesized below and further developed in participants’ working notes [20, 21, 22, 23]. Team IES [22] (Top1) utilized a Swin Transformer V2 Base [24] architecture as a feature extractor and used a similar approach for meta-data integration as Ren et al. [25] from the previous edition of FungiCLEF [13]. Besides, they introduced (i) a poisonous re-ranking that prevents predicting an edible species when there is a significant chance of the sample being poisonous, and (ii) a genus loss that improves the feature space’s regularization. 1 Each team usually had different "best" submissions for each track. Jack Etheredge [20] (Top2) combined visual information with metadata using MetaFormer-0 and MetaFormer-2 [26] and further improved the ensemble by a vision-only CAFormer-S18 [27], and proposed a novel application of openGAN [28] for open-set recognition of fine-grained images utilizing WGAN-GP [29]. Team upupup [21] (Top3) used Dynamic MLP [30] for the fusion of image features and metadata, identifying unknown classes using an entropy-based approach, training with a marginal expected loss for recognizing poisonous mushrooms while maintaining accuracy. Team DS@GT [23] (Top82 ) utilized DINOv2 visual embeddings [31] (namely the dinov2-large model with register for final submission) which were combined with metadata in a classifier head. The model was trained with a composite loss function, consisting of the Seesaw loss [32] and a binary cross entropy loss for poisonous species classification. 5. Conclusions and Discussion This paper presents an overview and results evaluation of the third edition of the FungiCLEF challenge, organized in conjunction with the CLEF LifeCLEF lab, and CVPR-FGVC11 — The Tenth Workshop on Fine-Grained Visual Categorization held within the CVPR conference. This challenge continues to push the boundaries of fine-grained visual categorization by bringing together diverse methodologies and innovative approaches from leading research teams worldwide. By introducing an updated dataset, emphasizing model efficiency, and addressing both species recognition and poisonous mushroom identification, FungiCLEF 2024 has fostered innovation and practical solutions for fungi recognition. The diverse approaches employed by the top-performing teams illustrate the evolving landscape of fungi recognition challenges. From sophisticated model architectures to novel techniques for handling unknown species and balancing species classification with poisonous species identification, participants showcased groundbreaking solutions. They built on the findings of previous challenges, particularly in the encoding of metadata, demonstrating continuous advancement in the field. The strict computational constraints imposed on submissions ensured that the resulting models were not only accurate but also practical for deployment in resource-limited environments and motivated participants to focus on principal improvements rather than training large ensembles of complex models. However, we also observed that enforcing the computational limits through a submission system caused additional technical difficulties with submission to some participants. Future editions shall thrive to further simplify the submission process for the participants. With the advances in recognition accuracy for well-known species, we propose that future work should focus on the more challenging cases, specifically few-shot classification techniques. This ap- proach, with its potential to push the fungal species recognition forward, would enable more robust identification of rare or newly discovered fungal species with limited training data. As the field pro- gresses, the ultimate goal remains to develop robust, accurate, and accessible fungi recognition systems that can support both expert mycologists and citizen scientists in documenting and understanding fungal biodiversity. Acknowledgments LP and JM were supported by the Technology Agency of the Czech Republic, project No. SS05010008 and project No. SS73020004. 2 This team encountered some issues while submitting to HuggingFace, but achieved better results than Top1 in their post- competition submissions. References [1] M. Sulc, L. Picek, J. Matas, T. Jeppesen, J. Heilmann-Clausen, Fungi recognition: A practical use case, in: Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, 2020, pp. 2316–2324. [2] L. Picek, M. Šulc, J. Matas, J. Heilmann-Clausen, T. S. Jeppesen, E. Lind, Automatic fungi recognition: Deep learning meets mycology, Sensors 22 (2022) 633. [3] T. G. Frøslev, J. Heilmann-Clausen, C. Lange, T. Læssøe, J. H. Petersen, U. Søchting, T. S. Jeppesen, J. Vesterholt, Danish mycological society, fungal records database (2019). [4] D. Taylor, T. Bruns, Community structure of ectomycorrhizal fungi in a pinus muricata forest: minimal overlap between the mature forest and resistant propagule communities, Molecular Ecology 8 (1999) 1837–1850. [5] L. Boddy, T. H. Jones, Interactions between basidiomycota and invertebrates, in: British mycological society symposia series, volume 28, Elsevier, 2008, pp. 155–179. [6] M. Sulc, L. Picek, J. Matas, Plant recognition by inception networks with test-time class prior estimation., in: CLEF (Working Notes), 2018. [7] K.-H. Liu, M.-H. Yang, S.-T. Huang, C. Lin, Plant species classification based on hyperspectral imaging via a lightweight convolutional neural network model, Frontiers in Plant Science 13 (2022) 855660. [8] L. Picek, M. Šulc, J. Matas, T. S. Jeppesen, J. Heilmann-Clausen, T. Læssøe, T. Frøslev, Danish fungi 2020-not just another image recognition dataset, in: Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, 2022, pp. 1525–1535. [9] A. Joly, L. Picek, S. Kahl, H. Goëau, V. Espitalier, C. Botella, B. Deneu, D. Marcos, J. Estopinan, C. Leblanc, T. Larcher, M. Šulc, M. Hrúz, M. Servajean, et al., Overview of LifeCLEF 2024: Challenges on species distribution prediction and identification, in: International Conference of the Cross-Language Evaluation Forum for European Languages, Springer, 2024. [10] L. Picek, M. Šulc, J. Matas, J. Heilmann-Clausen, Overview of fungiclef 2022: Fungi recognition as an open set classification problem, CEUR-WS, 2022. [11] L. Picek, M. Šulc, R. Chamidullin, J. Matas, Overview of fungiclef 2023: Fungi recognition beyond 1/0 cost, in: CLEF 2023-Conference and Labs of the Evaluation Forum, 2023. [12] L. Picek, M. Šulc, J. Heilmann-Clausen, J. Matas, Overview of FungiCLEF 2022: Fungi recognition as an open set classification problem, in: Working Notes of CLEF 2022 - Conference and Labs of the Evaluation Forum, 2022. [13] L. Picek, M. Šulc, J. Heilmann-Clausen, J. Matas, Overview of FungiCLEF 2023: Fungi recognition beyond 0-1 cost, in: Working Notes of CLEF 2023 - Conference and Labs of the Evaluation Forum, 2023. [14] T. Lorieul, E. Cole, B. Deneu, M. Servajean, A. Durso11, I. Bolon, H. Glotin, R. Planqué, R. R. de Castaneda, W.-P. Vellinga, et al., Overview of lifeclef 2021: An evaluation of machine-learning based species identification and species distribution prediction, in: Experimental IR Meets Multi- linguality, Multimodality, and Interaction: 12th International Conference of the CLEF Association, CLEF 2021, Virtual Event, September 21–24, 2021, Proceedings, volume 12880, Springer Nature, 2021, p. 371. [15] A. Joly, H. Goëau, S. Kahl, B. Deneu, M. Servajean, E. Cole, L. Picek, R. R. De Castaneda, I. Bolon, A. Durso, et al., Overview of lifeclef 2020: a system-oriented evaluation of automated species identification and species distribution prediction, in: International Conference of the Cross- Language Evaluation Forum for European Languages, Springer, 2020, pp. 342–363. [16] A. Joly, H. Goëau, S. Kahl, L. Picek, C. Botella, D. Marcos, M. Šulc, M. Hrúz, T. Lorieul, S. S. Moussi, et al., Lifeclef 2023 teaser: Species identification and prediction challenges, in: European Conference on Information Retrieval, Springer, 2023, pp. 568–576. [17] M. Tan, Q. Le, Efficientnet: Rethinking model scaling for convolutional neural networks, in: International conference on machine learning, PMLR, 2019, pp. 6105–6114. [18] A. Joly, L. Picek, S. Kahl, H. Goëau, V. Espitalier, C. Botella, B. Deneu, D. Marcos, J. Estopinan, C. Leblanc, et al., Lifeclef 2024 teaser: Challenges on species distribution prediction and identifica- tion, in: European Conference on Information Retrieval, Springer, 2024, pp. 19–27. [19] L. Picek, M. Hruz, A. M. Durso, Overview of SnakeCLEF 2024: Revisiting snake species identifica- tion in medically important scenarios, in: Working Notes of CLEF 2024 - Conference and Labs of the Evaluation Forum, 2024. [20] J. Etheredge, Openwgan-gp for fine-grained open-set fungi classification, in: Working Notes of CLEF 2024 - Conference and Labs of the Evaluation Forum, 2024. [21] B.-F. Tan, Y.-Y. Li, P. Wang, L. Zhao, X.-S. Wei, Say no to the poisonous fungi: An effective strategy for reducing 0-1 cost in fungiclef2024, in: Working Notes of CLEF 2024 - Conference and Labs of the Evaluation Forum, 2024. [22] S. Wolf, P. Thelen, J. Beyerer, Poison-aware open-set fungi classification: Reducing the risk of poisonous confusion, in: Working Notes of CLEF 2024 - Conference and Labs of the Evaluation Forum, 2024. [23] C. Chiu, M. Heil, T. Kim, A. Miyaguchi, Fine-grained classification for poisonous fungi identification with transfer learning, in: Working Notes of CLEF 2024 - Conference and Labs of the Evaluation Forum, 2024. [24] Z. Liu, H. Hu, Y. Lin, Z. Yao, Z. Xie, Y. Wei, J. Ning, Y. Cao, Z. Zhang, L. Dong, et al., Swin transformer v2: Scaling up capacity and resolution, in: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2022, pp. 12009–12019. [25] H. Ren, H. Jiang, W. Luo, M. Meng, T. Zhang, Entropy-guided open-set fine-grained fungi recognition, in: Working Notes of CLEF 2023 - Conference and Labs of the Evaluation Forum, 2023. [26] Q. Diao, Y. Jiang, B. Wen, J. Sun, Z. Yuan, Metaformer: A unified meta framework for fine-grained recognition, arXiv preprint arXiv:2203.02751 (2022). [27] W. Yu, C. Si, P. Zhou, M. Luo, Y. Zhou, J. Feng, S. Yan, X. Wang, Metaformer baselines for vision, IEEE Transactions on Pattern Analysis and Machine Intelligence (2023). [28] S. Kong, D. Ramanan, Opengan: Open-set recognition via open data generation, in: Proceedings of the IEEE/CVF International Conference on Computer Vision, 2021, pp. 813–822. [29] I. Gulrajani, F. Ahmed, M. Arjovsky, V. Dumoulin, A. C. Courville, Improved training of wasserstein gans, Advances in neural information processing systems 30 (2017). [30] L. Yang, X. Li, R. Song, B. Zhao, J. Tao, S. Zhou, J. Liang, J. Yang, Dynamic mlp for fine-grained image classification by leveraging geographical and temporal information, in: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2022, pp. 10945–10954. [31] M. Oquab, T. Darcet, T. Moutakanni, H. Vo, M. Szafraniec, V. Khalidov, P. Fernandez, D. Haziza, F. Massa, A. El-Nouby, et al., Dinov2: Learning robust visual features without supervision, arXiv preprint arXiv:2304.07193 (2023). [32] J. Wang, W. Zhang, Y. Zang, Y. Cao, J. Pang, T. Gong, K. Chen, Z. Liu, C. C. Loy, D. Lin, Seesaw loss for long-tailed instance segmentation, in: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 9695–9704.