<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Poison-Aware Open-Set Fungi Classification: Reducing the Risk of Poisonous Confusion</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Stefan</forename><surname>Wolf</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Vision and Fusion Lab</orgName>
								<orgName type="institution">Karlsruhe Institute of Technology KIT</orgName>
								<address>
									<addrLine>Vincenz-Prießnitz-Straße 3</addrLine>
									<postCode>76131</postCode>
									<settlement>Karlsruhe</settlement>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
							<affiliation key="aff1">
								<orgName type="department" key="dep1">Fraunhofer IOSB</orgName>
								<orgName type="department" key="dep2">Institute of Optronics</orgName>
								<orgName type="department" key="dep3">System Technologies and Image Exploitation</orgName>
								<address>
									<addrLine>Fraunhoferstrasse 1</addrLine>
									<postCode>76131</postCode>
									<settlement>Karlsruhe</settlement>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
							<affiliation key="aff2">
								<orgName type="department">Fraunhofer Center for Machine Learning</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Philipp</forename><surname>Thelen</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Vision and Fusion Lab</orgName>
								<orgName type="institution">Karlsruhe Institute of Technology KIT</orgName>
								<address>
									<addrLine>Vincenz-Prießnitz-Straße 3</addrLine>
									<postCode>76131</postCode>
									<settlement>Karlsruhe</settlement>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
							<affiliation key="aff1">
								<orgName type="department" key="dep1">Fraunhofer IOSB</orgName>
								<orgName type="department" key="dep2">Institute of Optronics</orgName>
								<orgName type="department" key="dep3">System Technologies and Image Exploitation</orgName>
								<address>
									<addrLine>Fraunhoferstrasse 1</addrLine>
									<postCode>76131</postCode>
									<settlement>Karlsruhe</settlement>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
							<affiliation key="aff2">
								<orgName type="department">Fraunhofer Center for Machine Learning</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Jürgen</forename><surname>Beyerer</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Vision and Fusion Lab</orgName>
								<orgName type="institution">Karlsruhe Institute of Technology KIT</orgName>
								<address>
									<addrLine>Vincenz-Prießnitz-Straße 3</addrLine>
									<postCode>76131</postCode>
									<settlement>Karlsruhe</settlement>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
							<affiliation key="aff1">
								<orgName type="department" key="dep1">Fraunhofer IOSB</orgName>
								<orgName type="department" key="dep2">Institute of Optronics</orgName>
								<orgName type="department" key="dep3">System Technologies and Image Exploitation</orgName>
								<address>
									<addrLine>Fraunhoferstrasse 1</addrLine>
									<postCode>76131</postCode>
									<settlement>Karlsruhe</settlement>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
							<affiliation key="aff2">
								<orgName type="department">Fraunhofer Center for Machine Learning</orgName>
							</affiliation>
						</author>
						<title level="a" type="main">Poison-Aware Open-Set Fungi Classification: Reducing the Risk of Poisonous Confusion</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">0B1BFEB8E67E0C9D2C88554441ABA921</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T17:55+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Fungi classification</term>
					<term>Open-set classification</term>
					<term>FungiCLEF</term>
					<term>Entropy</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>The FungiCLEF 2024 challenge aims to foster research in the field of application-oriented fine-grained open-set classification. Particularly, it sets the challenge to optimize fungi species classification while recognizing unknown species with the evaluation of multiple metrics targeting the problems of actual use-cases, e.g., the risk of a highly detrimental confusion of a poisonous species for an edible species. To develop a well-performing approach, we focus on reducing this particular risk by introducing multiple improvements. The major improvements are a poisonous reranking which prevents predicting an edible species while a significant chance of the sample being poisonous exists and a genus loss which provides additional training information improving the regularization of the feature space. The advancements provide a large improvement in terms of poisonous confusion but also in terms of overall classification accuracy. With this approach, we achieved the 1 st place in the challenge's main metric. Code is available at https://huggingface.co/stefanwolf/fungi2024.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Fine-grained open-set classification is an important topic in the biology context in order to find samples of rare species and to provide inexperienced citizen scientist with a support to identify species of plants and animals. Particularly, fungi species classification has an additional use case: identifying poisonous species to reduce the risk of accidental eating poisonous fungi. Thus, the FungiCLEF 2024 challenge <ref type="bibr" target="#b0">[1]</ref>, part of the LifeCLEF 2024 lab <ref type="bibr" target="#b1">[2]</ref>, sets up the task of open-set fungi species classification with an emphasis on correctly identifying fungi species while preventing confusing poisonous species for edible species.</p><p>While the task of open-set fungi classification has been intensively investigated in the recent iterations of the FungiCLEF challenge <ref type="bibr" target="#b2">[3,</ref><ref type="bibr" target="#b3">4]</ref>, optimizing the poisonous confusion error has only been lightly approached rendering the room for improvements quite large. Thus, we focus on the poisonous confusion error and can achieve significant gains with multiple advancements. These advancements are • a poisonous reranking which predicts the highest ranking poisonous species if its confidence is only lower by a certain factor than the overall highest ranking species. • a genus loss that regularizes the feature space by incorporating the genus label in training.</p><p>• a second open-set threshold to reduce the risk of misclassifying a poisonous sample as an unknown species. • a two-stage metadata integration that enhances the overall classification accuracy.</p><p>CLEF 2024: Conference and Labs of the Evaluation Forum, September 09-12, 2024, Grenoble, France stefan.wolf@kit.edu (S. Wolf); philipp.thelen@iosb.fraunhofer.de (P. Thelen); juergen.beyerer@iosb.fraunhofer.de (J. Beyerer)  </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Related work</head><p>A wide range of approaches have been proposed targeting fine-grained fungi classification in wildlife images. Sulc et. <ref type="bibr" target="#b4">[5]</ref> employ an ensemble of CNNs to classify images of fungi. Picek et al. <ref type="bibr" target="#b5">[6]</ref> propose a simple but effective probabilistic strategy to exploit metadata in order to improve the accuracy of fine-grained fungi classification. Kiss and Czùni <ref type="bibr" target="#b6">[7]</ref> provide a study about a broad-range of design choices optimizing mushroom type classification accuracy. The 2022 <ref type="bibr" target="#b2">[3]</ref> and the 2023 <ref type="bibr" target="#b3">[4]</ref> iterations of the FungiCLEF challenge summarize a variety of approaches with the 2022 iteration being focused on improving open-set fungi classification and the 2023 iteration emphasizing the importance of choosing metrics based on use cases to focus research on relevant aspects, e.g., focusing on reducing the confusion of poisonous species with edible species.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">FungiCLEF 2024 challenge</head><p>The 2024 iteration of the FungiCLEF challenge <ref type="bibr" target="#b0">[1]</ref> aims to stimulate research about efficient open-set fungi classification. The target is to distinguish 1,604 fungi species by using an observation consisting of one or multiple images with different perspectives and additional metadata information such as habitat, substrate, time and location. Apart from distinguishing the known species, the submitted approach needs to be able to solve an open-set scenario, i.e., it needs to recognize whether a sample is of a species not part of the training samples. The provided training data consists of the Danish Fungi 2020 dataset <ref type="bibr" target="#b5">[6]</ref>, the validation data consists of the test set of FungiCLEF's 2022 iteration <ref type="bibr" target="#b2">[3]</ref> and the test data consists of new data for the 2024 iteration. The evaluation is done based on three metrics:</p><p>• Track 1: Classification error -standard classification with "unknown" category. • Track 2: Poisonous confusion error -cost for confusing edible species for poisonous and vice versa (with 100× weight for confusing edible species for poisonous). • Track 3: User-focused error -user-focused loss composed of both the classification error and the poisonous/edible confusion.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Approach</head><p>Our challenge submission is based on the approach by Wolf and Beyerer <ref type="bibr" target="#b7">[8]</ref>. To simplify the training and due to insignificant impact, we refrain from applying the resampling-based class balancing. We extend the approach by applying several improvements as described in this section, i.e., a two-stage integration of metadata, a genus loss, a poisonous reranking and a two-threshold open-set recognition strategy. The overall architecture of our approach is shown in Figure <ref type="figure" target="#fig_0">1</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1.">Model architecture</head><p>We employ a Swin Transformer V2 Base <ref type="bibr" target="#b8">[9]</ref> as an image feature extractor backbone. Additionally, we use the metadata information provided by the Danish Fungi dataset <ref type="bibr" target="#b5">[6]</ref> to improve the classification accuracy. The metadata is encoded similarly to the approach by Ren et al. <ref type="bibr" target="#b9">[10]</ref> We encode the month 𝑚 and day 𝑑 of each observation as a vector</p><formula xml:id="formula_0">(︀ sin( 2𝜋𝑚 12 ), cos( 2𝜋𝑚 12 ), sin( 2𝜋𝑑 31 ), cos( 2𝜋𝑑<label>31</label></formula><p>) )︀ ⊺ . The geographical locations country code, substrate and habitat are encoded as one-hot vectors. All metadata vectors are concatenated and fed into two fully connected layers with an output size of 64 and each being followed by a ReLU activation and a layer norm <ref type="bibr" target="#b10">[11]</ref>. The resulting metadata feature vector and the image feature vector are concatenated and fed into a final linear classification layer followed by a softmax activation. We apply an auxiliary second classification head for predicting the genus of a sample during training which is fed the image features.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2.">Training process</head><p>For training the model, we apply two losses: a classification loss on the species level as commonly used and an auxiliary classification loss on the genus level. Both are label-smooth losses <ref type="bibr" target="#b11">[12]</ref> with a smoothing value of 0.9. To prevent a degradation of the image features when training with metadata, we use a two-stage training with the first stage only training the image classification stream of the network and the second stage training the complete network including the metadata feature extractor with a newly initialized species classifier.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.3.">Inference</head><p>During inference, we extract the feature vectors of all images of an observation and concatenate the observation-wise mean of the image features with the metadata features before feeding the result into the species classification head. Based on the resulting softmaxed confidence scores, we apply our poisonous reranking which reranks the poisonous species with the highest confidence to the top of the species ranking if its confidence is higher than the actual top-1 species' confidence divided by a poisonous reranking factor 𝛼. After the poisonous reranking, we apply the entropy-based open-set thresholding based on the approach by Ren et al. <ref type="bibr" target="#b9">[10]</ref> If the entropy of the output confidences is above a certain threshold 𝜏 , we predict the observation to be out-of-distribution. We extend this approach by employing two thresholds 𝜏 𝑒 and 𝜏 𝑝 , which are applied if the predicted species is edible or poisonous, respectively. The threshold 𝜏 𝑝 is selected higher than the threshold 𝜏 𝑒 to reduce the risk of misclassifying a posionous species as an out-of-distribution species. Both, the poisonous reranking and the second threshold, are improvements targeting the challenge's Track 2 metric which is measuring the poisonous confusion with a significantly higher weight for mispredicting poisonous species than edible species.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Evaluation</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.1.">Datasets</head><p>We use the official FungiCLEF 2024 <ref type="bibr" target="#b0">[1]</ref> datasets. The Danish Fungi 2020 <ref type="bibr" target="#b5">[6]</ref> dataset is used for training. All metrics reported in this study are based on the official validation set which is the test set of FungiCLEF's 2022 iteration <ref type="bibr" target="#b2">[3]</ref>. The test set for the official ranking is a set of images which has not been disclosed publicly prior to the end of the challenge. Only the results of the public part of the test set was publicly visible with the results on the private part only being disclosed after the challenge deadline.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.2.">Implementation and baseline</head><p>We use the MMPreTrain <ref type="bibr" target="#b12">[13]</ref> classification framework based on PyTorch <ref type="bibr" target="#b13">[14]</ref> for the training and the inference of the models. All models are pre-trained on the ImageNet-21k dataset <ref type="bibr" target="#b14">[15]</ref> and trained for 24 epochs with an AdamW optimizer <ref type="bibr" target="#b15">[16]</ref>, a base learning rate of 6.25 • 10 −5 , a learning rate warm-up for 2100 iterations and a cosine learning rate decay. We train with a total batch size of 128. The metadata training is performed for two epochs in a second stage with a frozen image encoder.</p><p>Our image pre-processing pipeline for training includes a random crop of an image area between 8% and 100%, a resize to 384×384 pixels, a random horizontal image flip, RandAugment <ref type="bibr" target="#b16">[17]</ref> and random erasing <ref type="bibr" target="#b17">[18]</ref>. We use 8 Nvidia A100 GPUs for training. The pre-processing pipeline for the inference includes an image scaling with 438 pixels output size on the shorter edge and a center crop of size 384×384 pixels.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.3.">Genus loss</head><p>We compare a training with and without our genus loss in Table <ref type="table" target="#tab_1">1</ref>. All metrics are improved by the application of the genus loss. Particularly, the Track 2 error focused on identifying poisonous samples as such improves significantly with a drop of 0.22 to 0.18. Nonetheless, also the classification-focused F1 score and Track 1 error show an improvement. The strong increase in terms of identifying poisonous species is likely due to most genus containing only edible species. Thus, considering the genus level in training the feature space results in a denser feature representation of these poisonous-wise uniform genus. Therefore, the chance of misclassification of a species of a uniformly edible genus with a nonedible species is heavily reduced. The improvements in terms of classification accuracy are probably induced by species with a low number of samples. The risk of misclassifying them with species from other genus due to a lack of variance in the data is reduced when also training the feature space on the genus level.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.4.">Poisonous reranking</head><p>We apply the poisonous reranking additional to the genus loss and evaluate it with different factors. It reranks poisonous species to the top-1 if their confidence is higher than the actual top-1's confidence divided by an 𝛼 &gt; 1 in order to prevent costly confusions of poisonous fungi with edible fungi. The results are shown in Table <ref type="table" target="#tab_2">2</ref>. As expected, it drastically reduces the Track 2 error from 0.181 to 0.076 due to a lower number of samples misclassified as an edible species. While the Track 1 error is slightly increased from 0.41 to 0.416 due to some so far correctly classified samples from edible species now being misclassified as a poisonous species, the increase is small compared to the drop in terms of Track 2 error resulting in a significant drop in the overall Track 3 error from 0.591 to 0.492. Out of the evaluated values of 𝛼, the Track 2 and 3 errors are reduced until an 𝛼 of 10 while a value of 20 is leading to an increase in all error metrics. Particularly, also the Track 2 error increases showing that even the comparatively lowly weighted case of mispredicting an edible species for a poisonous species now playing a significant role.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.5.">Open-set recognition</head><p>We employ the entropy-based open-set recognition by Ren et al. <ref type="bibr" target="#b9">[10]</ref> and extend it by a second threshold for poisonous species. If the predicted species is poisonous, a higher entropy is needed to classify the sample as out-of-distribution since out-of-distribution samples are considered edible by the Track 2 metric and thus, misclassifying a poisonous sample as out-of-distribution increases the Track 2 error heavily. We compare it to applying no open-set recognition and applying a simple softmax-based thresholding. The results including genus loss and poisonous reranking as baseline are shown in Table <ref type="table" target="#tab_3">3</ref> and indicate an improvement in all metrics for both open-set recognition methods. The entropy-based approach with two thresholds provide an additional improvement over the softmax-based thresholding.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.6.">Metadata</head><p>We integrate metadata information in the inference process by feeding the encoded metadata through two fully connected layers and concatenating the resulting vector to the feature vector of the image encoder before the final linear classification layer. The impact of this metadata exploitation strategy is shown in Table <ref type="table" target="#tab_4">4</ref> including all previously mentioned improvements. The results show a significant improvement across all evaluated metrics. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.7.">Final model</head><p>The final best-performing model includes all proposed improvements with the following adjustments:</p><p>1. a poisonous reranking factor 𝛼 of 13, the overall best performing value on the public test set.</p><p>2. an open-set entropy thresholds of 𝜏 𝑒 = 2.5 for edible species and 𝜏 𝑝 = 7 for poisonous species.</p><p>3. including the validation set in training with unknown samples being assigned a vector with each element having the same value as target label similar to the approach by Ren et al. <ref type="bibr" target="#b9">[10]</ref> and the true genus label for the genus loss.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.8.">Challenge results</head><p>The final private test set results of the top-5 challenge participants are shown in Table <ref type="table" target="#tab_5">5</ref>. We ranked first with the lowest error in the main metric Track 3 due to a high emphasis on optimizing Track 2 while not lacking too far behind in terms of Track 1. Particularly, we achieve the first place with an efficient solution that consists of only a single model. While the runner-up team achieved a better Track 1 error, the Track 2 error is almost twice as high. In contrast, the third-placed team achieved an even better Track 2 metric than our approach. However, this achievement comes at a large Track 1 error outweighing the advantage.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Conclusion</head><p>In this study, we described our top-ranking approach for the FungiCLEF 2024 challenge. With a high emphasis on reducing the risk of confusing poisonous species for edible species, we propose several advancements which improve the respective error drastically while also improving the overall classification accuracy. Particularly, we introduced a poisonous reranking, a genus loss, two-threshold open-set recognition and an efficient two-stage metdata exploitation strategy.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Overview of our approach. We employ a Swin Transformer V2 Base backbone to extract image features. The image features are passed to an auxiliary genus loss during training. The metadata features are processed by a multi-layer-perceptron and thereafter, they are combined with the image features and fed into the species classifier. Afterwards, our poisonous reranking and the two-threshold entropy-guided open-set recognition is applied.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 1</head><label>1</label><figDesc>Impact of the auxiliary genus loss. Our auxiliary genus loss shows a small improvement for the classic classification metrics F1 score and Track 1 (classification error). However, particularly the poison-focused Track 2 metric is drastically improved due to better feature separation on the genus level which is important for distinguishing poisonous from edible species. Note: the images of each observation are combined by a post-classification mean fusion.</figDesc><table><row><cell cols="5">Genus loss F1 score Track 1 Track 2 Track 3</cell></row><row><cell>No</cell><cell>49.1</cell><cell>0.413</cell><cell>0.220</cell><cell>0.633</cell></row><row><cell>Yes</cell><cell>49.7</cell><cell>0.410</cell><cell>0.181</cell><cell>0.591</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 2</head><label>2</label><figDesc>Impact of the poisonous reranking with different values of 𝛼. The results show the positive impact of reducing the risk of misclassifying a poisonous species for an edible species with significantly lowered Track 2 and 3 errors outweighing the small increase in terms of Track 1 error due to an overall increased species misclassification. Note: the images of each observation are combined by a post-classification mean fusion.</figDesc><table><row><cell cols="5">Poisonous reranking factor 𝛼 F1 score Track 1 Track 2 Track 3</cell></row><row><cell>-</cell><cell>49.7</cell><cell>0.410</cell><cell>0.181</cell><cell>0.591</cell></row><row><cell>2</cell><cell>49.7</cell><cell>0.411</cell><cell>0.144</cell><cell>0.555</cell></row><row><cell>5</cell><cell>49.3</cell><cell>0.413</cell><cell>0.106</cell><cell>0.518</cell></row><row><cell>10</cell><cell>49.1</cell><cell>0.416</cell><cell>0.076</cell><cell>0.492</cell></row><row><cell>20</cell><cell>48.8</cell><cell>0.421</cell><cell>0.079</cell><cell>0.500</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 3</head><label>3</label><figDesc>Impact of the two-threshold entropy-based open-set recognition. For the softmax-score-based recognition, we report results for thresholds optimized separately for each metric while we report results for a pair of thresholds for the entropy-based strategy which is manually optimized for Track 3. Thus, the softmax-score-based approach shows the best results for all metrics but the combined Track 3. For the Track 3 metric, the entropy-based approach outperforms the softmax score.</figDesc><table><row><cell>Open-set recognition</cell><cell cols="4">F1 score Track 1 Track 2 Track 3</cell></row><row><cell>-</cell><cell>49.9</cell><cell>0.412</cell><cell>0.082</cell><cell>0.494</cell></row><row><cell>Softmax score</cell><cell>52.1</cell><cell>0.364</cell><cell>0.070</cell><cell>0.469</cell></row><row><cell cols="2">Entropy-based with two thresholds 50.3</cell><cell>0.376</cell><cell>0.074</cell><cell>0.449</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_4"><head>Table 4</head><label>4</label><figDesc>Impact of the metadata integration. Exploiting the metadata gives a significant improvement for all metrics.</figDesc><table><row><cell cols="5">Metadata F1 score Track 1 Track 2 Track 3</cell></row><row><cell>No</cell><cell>50.3</cell><cell>0.378</cell><cell>0.068</cell><cell>0.445</cell></row><row><cell>Yes</cell><cell>56.4</cell><cell>0.329</cell><cell>0.051</cell><cell>0.380</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_5"><head>Table 5</head><label>5</label><figDesc>Overview of the top-5 challenge submissions. The best performing submission according to the challenge's main metric Track 3 is selected. Performance is measured on the private test set.</figDesc><table><row><cell>Team</cell><cell cols="4">F1 score Track 1 Track 2 Track 3</cell></row><row><cell>IES</cell><cell>54.3</cell><cell>0.311</cell><cell>0.090</cell><cell>0.401</cell></row><row><cell cols="2">jack-etheredge 54.9</cell><cell>0.244</cell><cell>0.163</cell><cell>0.407</cell></row><row><cell>upupup</cell><cell>53.6</cell><cell>0.390</cell><cell>0.072</cell><cell>0.462</cell></row><row><cell>chirmy</cell><cell>51.8</cell><cell>0.269</cell><cell>0.415</cell><cell>0.684</cell></row><row><cell>TingTing1999</cell><cell>51.4</cell><cell>0.275</cell><cell>0.438</cell><cell>0.713</cell></row></table></figure>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgments</head><p>This work was supported by the Helmholtz Association's Initiative and Networking Fund on the HAICORE@FZJ partition.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<author>
			<persName><forename type="first">L</forename><surname>Picek</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Sulc</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Matas</surname></persName>
		</author>
		<title level="m">Overview of FungiCLEF 2024: Revisiting fungi species recognition beyond 0-1 cost</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
	<note>Working Notes of CLEF 2024 -Conference and Labs of the Evaluation Forum</note>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Overview of lifeclef 2024: Challenges on species distribution prediction and identification</title>
		<author>
			<persName><forename type="first">A</forename><surname>Joly</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Picek</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Kahl</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Goëau</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Espitalier</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Botella</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Deneu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Marcos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Estopinan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Leblanc</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Larcher</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Šulc</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Hrúz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Servajean</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">International Conference of the Cross-Language Evaluation Forum for European Languages</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Overview of FungiCLEF 2022: Fungi recognition as an open set classification problem</title>
		<author>
			<persName><forename type="first">L</forename><surname>Picek</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Šulc</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Heilmann-Clausen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Matas</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Working Notes of CLEF 2022 -Conference and Labs of the Evaluation Forum</title>
				<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<monogr>
		<author>
			<persName><forename type="first">L</forename><surname>Picek</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Šulc</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Chamidullin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Matas</surname></persName>
		</author>
		<title level="m">Overview of fungiclef 2023: Fungi recognition beyond 1/0 cost</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
	<note>CLEF 2023-Conference and Labs of the Evaluation Forum</note>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Fungi recognition: A practical use case</title>
		<author>
			<persName><forename type="first">M</forename><surname>Sulc</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Picek</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Matas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Jeppesen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Heilmann-Clausen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision</title>
				<meeting>the IEEE/CVF Winter Conference on Applications of Computer Vision</meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="2316" to="2324" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Danish fungi 2020-not just another image recognition dataset</title>
		<author>
			<persName><forename type="first">L</forename><surname>Picek</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Šulc</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Matas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">S</forename><surname>Jeppesen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Heilmann-Clausen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Laessøe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Frøslev</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision</title>
				<meeting>the IEEE/CVF Winter Conference on Applications of Computer Vision</meeting>
		<imprint>
			<date type="published" when="2022">2022</date>
			<biblScope unit="page" from="1525" to="1535" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Mushroom image classification with cnns: A case-study of different learning strategies</title>
		<author>
			<persName><forename type="first">N</forename><surname>Kiss</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Czùni</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">12th International Symposium on Image and Signal Processing and Analysis (ISPA)</title>
				<imprint>
			<publisher>IEEE</publisher>
			<date type="published" when="2021">2021. 2021</date>
			<biblScope unit="page" from="165" to="170" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Optimizing fine-grained fungi classification for diverse application-oriented open-set metrics</title>
		<author>
			<persName><forename type="first">S</forename><surname>Wolf</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Beyerer</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Working Notes of the Conference and Labs of the Evaluation Forum</title>
				<meeting><address><addrLine>CLEF</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2023">2023. 2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Swin transformer v2: Scaling up capacity and resolution</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Hu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Lin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Yao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Xie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Wei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Ning</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Cao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Dong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Wei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Guo</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)</title>
				<meeting>the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)</meeting>
		<imprint>
			<date type="published" when="2022">2022</date>
			<biblScope unit="page" from="12009" to="12019" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Entropy-guided open-set fine-grained fungi recognition</title>
		<author>
			<persName><forename type="first">H</forename><surname>Ren</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Jiang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Luo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Meng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Zhang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Working Notes of the Conference and Labs of the Evaluation Forum</title>
				<meeting><address><addrLine>CLEF</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2023">2023. 2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<monogr>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">L</forename><surname>Ba</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">R</forename><surname>Kiros</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">E</forename><surname>Hinton</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1607.06450</idno>
		<title level="m">Layer normalization</title>
				<imprint>
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Rethinking the inception architecture for computer vision</title>
		<author>
			<persName><forename type="first">C</forename><surname>Szegedy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Vanhoucke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Ioffe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Shlens</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Wojna</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the IEEE conference on computer vision and pattern recognition</title>
				<meeting>the IEEE conference on computer vision and pattern recognition</meeting>
		<imprint>
			<date type="published" when="2016">2016</date>
			<biblScope unit="page" from="2818" to="2826" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<monogr>
		<ptr target="https://github.com/open-mmlab/mmpretrain" />
		<title level="m">Openmmlab&apos;s pre-training toolbox and benchmark</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
	<note>MMPreTrain Contributors</note>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Pytorch: An imperative style, high-performance deep learning library</title>
		<author>
			<persName><forename type="first">A</forename><surname>Paszke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Gross</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Massa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Lerer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Bradbury</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Chanan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Killeen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Lin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Gimelshein</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Antiga</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Advances in neural information processing systems</title>
		<imprint>
			<biblScope unit="volume">32</biblScope>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Imagenet: A large-scale hierarchical image database</title>
		<author>
			<persName><forename type="first">J</forename><surname>Deng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Dong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Socher</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L.-J</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Fei-Fei</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">IEEE conference on computer vision and pattern recognition</title>
				<imprint>
			<publisher>Ieee</publisher>
			<date type="published" when="2009">2009. 2009</date>
			<biblScope unit="page" from="248" to="255" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<monogr>
		<author>
			<persName><forename type="first">I</forename><surname>Loshchilov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Hutter</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1711.05101</idno>
		<title level="m">Decoupled weight decay regularization</title>
				<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Randaugment: Practical automated data augmentation with a reduced search space</title>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">D</forename><surname>Cubuk</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Zoph</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Shlens</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><forename type="middle">V</forename><surname>Le</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops</title>
				<meeting>the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops</meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Random erasing data augmentation</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Zhong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Zheng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Kang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Yang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the AAAI conference on artificial intelligence</title>
				<meeting>the AAAI conference on artificial intelligence</meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="volume">34</biblScope>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
