<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>AI Multimedia Lab at ImageCLEFmedical GANs 2025: Identifying Real-Image Usage in Generated Medical Images</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Alexandra-Georgiana Andrei</string-name>
          <email>alexandra.andrei@upb.ro</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Mihai Gabriel Constantin</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Mihai Dogariu</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Liviu-Daniel Stefan</string-name>
          <email>liviu_daniel.stefan@upb.ro</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Bogdan Ionescu</string-name>
          <email>bogdan.ionescu@upb.ro</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>AI Multimedia Lab, National University of Science and Technology Politehnica Bucharest</institution>
          ,
          <country country="RO">Romania</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <abstract>
        <p>This paper presents the participation of AI Multimedia Lab in the third edition of the 2025 ImageCLEFmedical GANs task, which investigates privacy and security concerns around generated synthetic medical images. This edition, the challenge comprises two complementary subtasks: (1) Detect which real images were used to train a GAN based on given synthetic outputs and (2) Attribute each synthetic image to its specific real-image subset of origin. We present our team's approach, which combines traditional deep learning techniques on two-step pipeline - feature extraction followed by clustering - for subtask 2, and Siamese Neural Networks for both subtasks. Evaluated on benchmark testing datasets of real and synthetic lung CT slices, our Siamese-based method achieved a Cohen's kappa of 0.036 for Subtask 1 and 99.04% accuracy for Subtask 2. Finally, we discuss the strengths and limitations of our methods and outline directions for improving the detection of training-data “fingerprints” in GAN-generated medical images. All code used in this study is available in Github 1.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;synthetic medical data</kwd>
        <kwd>Generative Adversarial Networks</kwd>
        <kwd>data augmentation</kwd>
        <kwd>ImageCLEFmedical GANs</kwd>
        <kwd>ImageCLEFbenchmarking lab</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Building on the foundation laid by previous editions [
        <xref ref-type="bibr" rid="ref1 ref2">1, 2</xref>
        ], part of ImageCLEF evaluation campaign [
        <xref ref-type="bibr" rid="ref3 ref4 ref5">3, 4,
5</xref>
        ], the third edition of the 2025 ImageCLEFmedical GANs [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] task continues to explore the intersection
of generative models and medical imaging, with a particular focus on data privacy and security. This
edition introduces two complementary subtasks: (1) detecting which real images were used to train a
generative model based on its synthetic outputs, and (2) identifying the specific real-image subset from
which a given synthetic image was generated. These subtasks aim to investigate whether synthetic
medical images carry identifiable traces - “fingerprints” - of the data they were trained on, raising
important implications for privacy-preserving data generation.
      </p>
      <p>This paper presents the participation of the AI Multimedia Lab – task organizing team – in the third
edition of the ImageCLEFmedical GANs task. The paper is structured as follows: Section 2 presents
the tasks and the datasets, Section 3 presents the proposed methods and the results are presented and
discussed in Section 4. Finally, the paper closes with Section 5, where we present the conclusions.</p>
    </sec>
    <sec id="sec-2">
      <title>2. The 2025 ImageCLEFmedical GANs Task</title>
      <sec id="sec-2-1">
        <title>2.1. Subtask 1 – “Detect Training Data Usage”</title>
        <p>This task introduced in the first edition has been continued in both following editions in the same setup.
The objective of the task is to detect “fingerprints” within the generated images to determine which of
the real images were used to train the GAN that generated the synthetic images. The data set provided
for this task consists of real and synthetic images generated, as described in Table 1.</p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Subtask 2 – “Identify Training Data Subsets”</title>
        <p>To address Subtask 2, the goal was to predict which training subset was used to generate each image in
the testing set. The training dataset was structured into two main folders: one containing real images
and another containing synthetic (generated) images as described in Table 2. Each of these folders had
ifve subfolders (t1–t5 for real, and gen_t1–gen_t5 for generated), where each pair of corresponding
subfolders shared a one-to-one mapping. For example, real images in t1 were used to generate synthetic
images in gen_t1. The testing dataset comprised 25,000 generated images, each of which needed to be
assigned to one of the five original real-image subsets.</p>
        <p>
          More information about both subtasks are available in the overview paper of the task [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ].
        </p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Proposed Methods</title>
      <sec id="sec-3-1">
        <title>3.1. Subtask 2 – “Identify Training Data Subsets”</title>
        <p>We started with the second subtask “Identify Training Data Subsets” for which we proposed two
diferent approaches:
• Two step pipeline: feature extraction using pre-trained models and classification using Support</p>
        <p>Vector Machine (SVM) and clustering using k-means and agglomerative hierarchical clustering;
• Siamese Neural Networks trained on (real, generated) pairs.</p>
        <sec id="sec-3-1-1">
          <title>3.1.1. Method 1 – Feature Clustering</title>
          <p>We adopted the pipeline presented by our team in the previous editions [7, 8] to the current task setup.
In this edition, we tested diferent variations of the pipeline presented in Figure 1 for detecting the real
subset of training data used to generate each of the synthetic images.</p>
          <p>Feature extraction – we employed four diferent pretrained models that were originally trained on
ImageNet dataset [9]: i) MobileNetV2 [10], ii) ResNet50 [11], iii) EficientNet [ 12], and iv) DenseNet [13].</p>
          <p>Clustering – Since the dataset includes five known training subsets, we explored three diferent
approaches to group and classify the data: i) k-means, ii) hierarchical clustering and iii) SVM. K-means
and hierarchical clustering are both unsupervised methods that group similar data points, but they work
in diferent ways: k-means clusters data by minimizing distance to cluster centers, while hierarchical
clustering builds a hierarchy based on how close points are to each other. SVM, on the other hand,
is a supervised method that learns to separate data based on labeled examples. By applying all three
methods, we aim to understand whether the extracted features can meaningfully represent the data.</p>
          <p>We kept 10% of the training set for the validation of the methods, and the rest of the data provided in
the training set was used to train the models in order to bring the pretrained models to train them on
our problem.</p>
        </sec>
        <sec id="sec-3-1-2">
          <title>3.1.2. Method 2 – Siamese Neural Networks</title>
          <p>A Siamese neural network [14] is specifically designed to learn similarity between pairs of inputs.
It consists of two identical subnetworks (or branches) that share the same weights and parameters.
Each subnetwork receives one of the input samples and transforms it into a feature representation,
or embedding. These embeddings are then compared using a distance metric such as the Euclidean
distance to determine how similar the inputs are. The network is typically trained with contrastive loss,
which minimizes the distance for similar pairs and maximizes it for dissimilar ones.</p>
          <p>
            We adopted a Siamese neural network architecture trained with contrastive loss to identify which
subset of the training data was used to generate each synthetic image in the testing set. This approach
enabled the model to learn a discriminative embedding space where real and synthetic images originating
from the same training subset are mapped close together, while those from diferent subsets are pushed
farther apart. The input pairs consisted of one real image and one synthetic image. Positive pairs were
formed from images originating from the same training subset, while negative pairs came from diferent
subsets. Each image was preprocessed by and normalizing pixel intensities to the [
            <xref ref-type="bibr" rid="ref1">0, 1</xref>
            ] range. These
pairs were used to train the Siamese model in diferent experimental setups.
          </p>
          <p>The structure of the network that was used for the results presented in this paper is depicted
in Figure 2 and it was trained over 30 epochs using the Adam optimizer and contrastive loss. The
architecture of each Siamese branch included a series of convolutional layers followed by max-pooling,
culminating in a dense layer with L2 normalization to produce compact embeddings. Specifically, the
network consisted of four convolutional blocks (with 16, 32, 64, and 128 filters), each followed by
ReLU activation and max-pooling, and a final dense layer of 256 units with L2 normalization. The
ifnal step was a distance computation layer, which calculated the Euclidean (L2) distance between the
two embeddings. Validation was performed using a set of positive and negative pairs to ensure that
each class was represented. After training, similarity scores were thresholded to compute validation
accuracy, distinguishing whether image pairs originated from the same training subset.</p>
          <p>For inference on the testing set, we first computed class centroids by averaging the embeddings of
real training images for each subset. Each testing image was passed through the embedding network to
generate its feature representation, which was then compared to all centroids using Euclidean distance.
The label of the closest centroid was assigned as the predicted subset for the testing image.</p>
        </sec>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Subtask 1 – “Detect Training Data Usage”</title>
        <p>We applied the same Siamese neural network architecture used in subtask 2 shown in Figure 2 to address
subtask 1. In this setup, both branches of the network received synthetic images from the training
dataset. For positive pairs, each synthetic image was matched with a real image that had been used to
train the generative model. For negative pairs, the synthetic image was paired with a real image that
was not part of the training data for that specific subset.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Results and Discussion</title>
      <sec id="sec-4-1">
        <title>4.1. Subtask 1 – “Detect Training Data Usage”</title>
        <p>Training dataset</p>
        <p>We evaluated the proposed Siamese Neural Network in two diferent training configurations and
obtained Cohen’s kappa scores of 0.35 and 0.50, respectively. These scores indicate a fair to moderate
agreement with the ground truth in identifying whether a given real image was used to train the
generative model. While not indicative of high precision, these results suggest that the model is able to
detect subtle patterns or “fingerprints” left by the training data in the synthetic images. A kappa of 0.50,
in particular, reflects a meaningful level of distinction between used and unused training data, although
there is still room for improvement in terms of discriminative reliability.</p>
        <p>Testing dataset</p>
        <p>Table 3 presents the results of two configurations of the proposed Siamese NN evaluated on the
testing dataset for subtask. The best-performing run (ID 1696) achieved a Cohen’s kappa of 0.036
with an accuracy of 51.80%, indicating only marginally better-than-random performance. The second
configuration (ID 1492) resulted in a negative kappa score (-0.044), suggesting performance slightly
worse than chance.</p>
        <p>While both models yielded similar F1-scores (∼ 0.54), the low kappa values reflect a weak agreement
with the ground truth and suggest that the network struggles to consistently distinguish between
images that were used in training and those that were not. This outcome highlights the challenging
nature of subtask 1, where the distinction between used and unused training images is subtle and
may not be reliably captured through visual similarity alone. These results point to the need for more
sensitive representations or alternative model strategies capable of detecting the latent “fingerprints”
left by GAN training processes.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Subtask 2 – “Identify Training Data Subsets”</title>
        <p>This section presents the results obtained using both proposed methods to address this subtask.
Table 4 summarizes the performance on the training dataset using feature clustering approaches.
Based on the validation accuracy, evaluated on a held-out portion of the training data. Among the
three classification strategies, SVM consistently achieved the highest validation accuracy across all
feature types, with ResNet50 + SVM yielding the best overall performance (52.59%). This suggests that
supervised classification is more efective in capturing the discriminative patterns embedded in the
pretrained features, while unsupervised clustering methods were less reliable for this task.</p>
        <p>To further analyze the classification behavior of the SVM with diferent feature representations,
we include confusion matrices (Figures 3a, 3b, 3c, and 3d) for the top-performing configurations:
DenseNet + SVM, ResNet50 + SVM, EficientNet + SVM, and MobileNetV2 + SVM. These matrices
provide insight into class-wise performance and highlight specific confusion patterns. DenseNet +
SVM (Figure 3a): The classifier performed strongly for class t3, with 1,573 correct predictions, and
showed good performance for t1 as well. However, misclassifications were observed: a large number
of t4 samples were predicted as t3 (576), and t2 samples as t1 (335). ResNet50 + SVM (Figure 3b): This
configuration yielded the most balanced performance across all classes. It achieved high accuracy for
t1 (1,292), t2 (1,070), and t3 (1,041), with relatively lower of-diagonal confusion compared to other
models. Some confusion persisted between t4 and both t2 and t3, but to a lesser extent than in other
configurations. The diagonal dominance in the confusion matrix suggests that ResNet50 features
are more discriminative and lead to more consistent classification. EficientNetB0 + SVM (Figure ??):
EficientNetB0 showed high accuracy for t1 and t3 (1,364 and 1,164 respectively), but was less consistent
for t4, where 706 samples were incorrectly predicted as t1. There was also mild confusion between
t2 and t1. The model seems to capture certain class-specific features well but lacks robustness for
classes with overlapping characteristics, particularly those positioned in the middle of the label set.
MobileNetV2 + SVM (Figure 3d): MobileNetV2 demonstrated generally stable performance across all
classes, with good classification of t3 (1,379 correct). However, t2 and showed notable confusion with t1
and t3.</p>
        <p>In order to visually analyze the clusters obtained with k-means and hierarchical clustering methods,
we plotted the PCA-2D projections for both real and synthetic images from the training dataset.</p>
        <p>Figures 4–7 display the features projected into a 2D space using PCA with two components. These
visualizations correspond to diferent feature extractors (MobileNetV2, ResNet-50, DenseNet-121, and
EficientNet) and provide insight into how well the clustering algorithms separate and group data.</p>
        <p>By comparing these figures alongside the quantitative metrics in Table 4, we observe that the overall
cluster structures produced by k-means and hierarchical clusteringare generally similar, confirming the
stability of clustering results across methods. Figure 4–a and 4–b show the PCA-2D projections using
the features extracted using DenseNet-121 network. Both methods show a broader spread in the PCA
space, especially along the first principal component. Despite this dispersion, both clustering algorithms
continue to group real and synthetic samples in a similar manner. Figures 5–a and 5–b illustrate the
clustering results based on EficientNet features. These projections show high overlap between real and
synthetic samples across most clusters. The results obtained using MobileNetV2, depicted in Figures 6–a
and 6–b show the PCA-2D projections using features extracted from MobileNetV2. Both clustering
methods yield clearly defined cluster boundaries, with a moderate degree of overlap between real and
synthetic images. The visual coherence of clusters across both methods suggests MobileNetV2 provides
a relatively disentangled feature space for both domains. Finally, Figures 7–a and 7–b illustrate the
clustering results based on ResNet-50 features. Both clustering methods capture similar structural
divisions in the embedding space, though the k-means clusters appear slightly more isotropic and
evenly distributed. Overall, these visual analyses support the conclusion that both clustering methods
produce consistent structures across diferent feature extractors. Moreover, they confirm that synthetic
data is generally well aligned with real data in the learned feature spaces.</p>
        <p>These results guided our selection of runs submitted – we selected only the best-performing
configurations (IDs 3, 6, 9, and 12) for oficial submission.</p>
        <p>When it comes to Siamese neural network described in Section 3, we obtained an accuracy of 100%.
Testing dataset</p>
        <p>Table 5 reports the performance of the proposed methods on the oficial testing dataset for subtask 2.
Among all evaluated approaches, the Siamese Neural Network (ID 1396) achieved a significantly higher
performance than all feature clustering methods, reaching an accuracy of 99.04%, with equally high
precision, recall, F1 score, and a specificity of 99.72%. This confirms the efectiveness of the Siamese
architecture in learning fine-grained similarity patterns between real and synthetic images.In contrast, the
feature-based SVM classifiers, while substantially less accurate, showed varying degrees of performance
depending on the feature extractor used. ResNet + SVM (ID 1268) achieved the best accuracy among
the clustering-based methods (52.36%), followed by EficientNet and DenseNet configurations. Notably,
MobileNetV2 + SVM lagged behind with the lowest performance (41.11% accuracy), indicating that its
learned features may be less separable for this task. Precision and specificity trends followed similar
patterns, reinforcing that deeper architectures such as ResNet and DenseNet yield more discriminative
embeddings.</p>
        <p>Overall, these results confirm that supervised similarity learning using a Siamese architecture
substantially outperforms unsupervised and supervised clustering approaches for identifying the origin of
synthetic medical images.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusions</title>
      <p>In this paper, we presented the participation of the AI Multimedia Lab in the third edition of the
ImageCLEFmedical GANs 2025 task, which focuses on identifying traces of real-image usage in synthetic
medical images generated by GANs. Our contributions addressed both subtasks of the challenge using
a combination of traditional machine learning and deep learning techniques.</p>
      <p>For subtask 1, which aimed to detect whether a real image was used in the training of a GAN, we
applied a Siamese neural network trained to diferentiate used versus unused real images based on
paired similarity with synthetic outputs. Although the training results suggested the model could
identify subtle training data patterns (kappa up to 0.50), performance on the oficial test set was limited
(a kappa score of 0.036), highlighting the dificulty of the task and the need for more sensitive modeling
approaches.</p>
      <p>In subtask 2, which required identifying the specific real-image subset used to generate each synthetic
image, we proposed two complementary approaches: (1) a feature clustering pipeline using pretrained
models and traditional classifiers, and (2) a Siamese neural network trained to learn subset-specific
embeddings. Among the evaluated methods, the Siamese network achieved the best performance by far,
with an accuracy of 99.04% on the test set, demonstrating its efectiveness in learning discriminative
patterns for subset attribution. Feature-based SVM classifiers performed reasonably well, with ResNet50
features achieving over 52% accuracy, but were clearly outperformed by the Siamese approach.</p>
      <p>Overall, our results show that while detecting whether an image was used in GAN training remains
a challenging problem, identifying the origin subset of synthetic images is feasible with high accuracy
using supervised similarity learning.</p>
    </sec>
    <sec id="sec-6">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the authors used ChatGPT-4o in order to: Grammar and spelling
check and improve writing style. After using this tool, the authors reviewed and edited the content as
needed and take full responsibility for the publication’s content.</p>
    </sec>
    <sec id="sec-7">
      <title>6. Acknowledgments</title>
      <p>The work of Alexandra Andrei was supported by a grant of the Ministry of Research, Innovation and
Digitization, CCCDI - UEFISCDI, project number PN-IV-P6-6.3-SOL-2024-0049, within PNCDI IV. The
work of Mihai Gabriel Constantin was supported by a grant of the Ministry of Research, Innovation
and Digitization, CCCDI - UEFISCDI, project number PN-IV-P6-6.3-SOL-2024-0060, within PNCDI IV.
[7] A. Andrei, M. G. Constantin, M. Dogariu, B. Ionescu, Ai multimedia lab at imageclefmedical gans
2024: Deep learning approaches for analyzing synthetic medical images, in: CLEF2024 Working
Notes, CEUR Workshop Proceedings, Grenoble, France, 2024.
[8] A.-G. Andrei, B. Ionescu, Aimultimedialab at imageclefmedical gans 2023: Determining"
fingerprints" of training data in generated synthetic images., in: CLEF (Working Notes), 2023, pp.
1379–1386.
[9] J. Deng, W. Dong, R. Socher, L.-J. Li, K. Li, L. Fei-Fei, Imagenet: A large-scale hierarchical image
database, in: 2009 IEEE Conference on Computer Vision and Pattern Recognition, 2009, pp.
248–255. doi:10.1109/CVPR.2009.5206848.
[10] M. Sandler, A. Howard, M. Zhu, A. Zhmoginov, L.-C. Chen, Mobilenetv2: Inverted residuals
and linear bottlenecks, in: Proceedings of the IEEE conference on computer vision and pattern
recognition, 2018, pp. 4510–4520.
[11] B. Koonce, B. Koonce, Resnet 50, Convolutional neural networks with swift for tensorflow: image
recognition and dataset categorization (2021) 63–72.
[12] B. Koonce, B. Koonce, Eficientnet, Convolutional neural networks with swift for Tensorflow:
image recognition and dataset categorization (2021) 109–123.
[13] F. Iandola, M. Moskewicz, S. Karayev, R. Girshick, T. Darrell, K. Keutzer, Densenet: Implementing
eficient convnet descriptor pyramids, arXiv preprint arXiv:1404.1869 (2014).
[14] D. Chicco, Siamese neural networks: An overview, Artificial neural networks (2021) 73–94.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>A.-G.</given-names>
            <surname>Andrei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Radzhabov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I.</given-names>
            <surname>Coman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Kovalev</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Ionescu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Müller</surname>
          </string-name>
          ,
          <article-title>Overview of imageclefmedical gans 2023 task: identifying training data “fingerprints” in synthetic biomedical images generated by gans for medical image security</article-title>
          ,
          <source>in: Working Notes of the Conference and Labs of the Evaluation Forum (CLEF</source>
          <year>2023</year>
          ), volume
          <volume>3497</volume>
          ,
          <year>2023</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>A.</given-names>
            <surname>Andrei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Radzhabov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Karpenka</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Prokopchuk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Kovalev</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Ionescu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Müller</surname>
          </string-name>
          ,
          <article-title>Overview of 2024 imageclefmedical gans task-investigating generative models' impact on biomedical synthetic images</article-title>
          ,
          <source>in: CLEF2024 Working Notes, CEUR Workshop Proceedings</source>
          , CEUR-WS. org, Grenoble, France,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>B.</given-names>
            <surname>Ionescu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Müller</surname>
          </string-name>
          ,
          <string-name>
            <surname>A.-M. Drăgulinescu</surname>
            ,
            <given-names>W.-W.</given-names>
          </string-name>
          <string-name>
            <surname>Yim</surname>
            ,
            <given-names>A. Ben</given-names>
          </string-name>
          <string-name>
            <surname>Abacha</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Snider</surname>
            , G. Adams,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Yetisgen</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Rückert</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>García Seco de Herrera</surname>
          </string-name>
          , et al.,
          <article-title>Overview of the imageclef 2023: multimedia retrieval in medical, social media and internet applications</article-title>
          ,
          <source>in: International Conference of the Cross-Language Evaluation Forum for European Languages</source>
          , Springer,
          <year>2023</year>
          , pp.
          <fpage>370</fpage>
          -
          <lpage>396</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>B.</given-names>
            <surname>Ionescu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Müller</surname>
          </string-name>
          ,
          <string-name>
            <surname>A.-M. Drăgulinescu</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Rückert</surname>
            ,
            <given-names>A. Ben</given-names>
          </string-name>
          <string-name>
            <surname>Abacha</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>García Seco de Herrera</surname>
            , L. Bloch,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Brüngel</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Idrissi-Yaghir</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Schäfer</surname>
          </string-name>
          , et al.,
          <source>Overview of the imageclef</source>
          <year>2024</year>
          :
          <article-title>Multimedia retrieval in medical applications</article-title>
          ,
          <source>in: International Conference of the Cross-Language Evaluation Forum for European Languages</source>
          , Springer,
          <year>2024</year>
          , pp.
          <fpage>140</fpage>
          -
          <lpage>164</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>B.</given-names>
            <surname>Ionescu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Müller</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.-C.</given-names>
            <surname>Stanciu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.-G.</given-names>
            <surname>Andrei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Radzhabov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Prokopchuk</surname>
          </string-name>
          , Ştefan, LiviuDaniel, M.-G. Constantin,
          <string-name>
            <given-names>M.</given-names>
            <surname>Dogariu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Kovalev</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Damm</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Rückert</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. Ben</given-names>
            <surname>Abacha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>García Seco de Herrera</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. M.</given-names>
            <surname>Friedrich</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Bloch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Brüngel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Idrissi-Yaghir</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Schäfer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. S.</given-names>
            <surname>Schmidt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T. M. G.</given-names>
            <surname>Pakull</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Bracke</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Pelka</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Eryilmaz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Becker</surname>
          </string-name>
          , W.-W. Yim,
          <string-name>
            <given-names>N.</given-names>
            <surname>Codella</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. A.</given-names>
            <surname>Novoa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Malvehy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Dimitrov</surname>
          </string-name>
          ,
          <string-name>
            <surname>R. J. Das</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          <string-name>
            <surname>Xie</surname>
            ,
            <given-names>H. M.</given-names>
          </string-name>
          <string-name>
            <surname>Shan</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Nakov</surname>
            , I. Koychev,
            <given-names>S. A.</given-names>
          </string-name>
          <string-name>
            <surname>Hicks</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Gautam</surname>
            ,
            <given-names>M. A.</given-names>
          </string-name>
          <string-name>
            <surname>Riegler</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          <string-name>
            <surname>Thambawita</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Halvorsen</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Fabre</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Macaire</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Lecouteux</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Schwab</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Potthast</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Heinrich</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Kiesel</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Wolter</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Stein</surname>
          </string-name>
          , Overview of imageclef 2025:
          <article-title>Multimedia retrieval in medical, social media and content recommendation applications, in: Experimental IR Meets Multilinguality</article-title>
          , Multimodality, and
          <string-name>
            <surname>Interaction</surname>
          </string-name>
          ,
          <source>Proceedings of the 16th International Conference of the CLEF Association (CLEF</source>
          <year>2025</year>
          ), Springer Lecture Notes in Computer Science LNCS, Madrid, Spain,
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>A.-G.</given-names>
            <surname>Andrei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. G.</given-names>
            <surname>Constantin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Dogariu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Radzhabov</surname>
          </string-name>
          , L.
          <string-name>
            <surname>-D. Ştefan</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>Prokopchuk</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          <string-name>
            <surname>Kovalev</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Müller</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Ionescu</surname>
          </string-name>
          ,
          <article-title>Overview of imageclefmedical 2025 GANs task: Training data analysis and ifngerprint detection</article-title>
          ,
          <source>in: CLEF2025 Working Notes, CEUR Workshop Proceedings</source>
          , CEUR-WS.org, Madrid, Spain,
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>