<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Overview of ImageCLEFMedical 2025 GANs Task: Training Data Analysis and Fingerprint Detection</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Alexandra-Georgiana Andrei</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Mihai Gabriel Constantin</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Mihai Dogariu</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ahmedkhan Radzhabov</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Liviu-Daniel S</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>tefan</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Yuri Prokopchuk</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Vassili Kovalev</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Henning Müller</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Bogdan Ionescu</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>AI Multimedia Lab, National University of Science and Technology Politehnica Bucharest</institution>
          ,
          <country country="RO">Romania</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Belarusian Academy of Sciences</institution>
          ,
          <addr-line>Minsk</addr-line>
          ,
          <country country="BY">Belarus</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>University of Applied Sciences Western Switzerland (HES-SO)</institution>
          ,
          <addr-line>Sierre</addr-line>
          ,
          <country country="CH">Switzerland</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>1873</year>
      </pub-date>
      <abstract>
        <p>The 2025 ImageCLEFmedical GANs Task - Controlling the Quality of Synthetic Medical Images created via GANs, continuing to investigate privacy and security concerns around using patient data to generate synthetic medical images. It comprises two complementary sub-tasks: the first extends prior editions by asking participants to detect which real images were used in training a Generative Adversarial Network to produce given synthetic outputs; the second builds on the 2024 findings by requiring teams to attribute each synthetic image to its specific real-image subset of origin. Ground-truth annotations and benchmark datasets of real and GAN-generated lung CT slices are provided for both tasks, and evaluation is based on Cohen's Kappa for Subtask 1 and accuracy for Subtask 2. 14 teams submitted runs for Subtask1 and 4 teams submitted runs for Subtask 2, totaling 95 submitted runs that used a variety of methods. This paper presents an overview of the task setup, datasets, and evaluation metrics, and summarizes and discusses the approaches and results of the .</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;generative models</kwd>
        <kwd>Generative Adversarial Networks</kwd>
        <kwd>medical synthetic data</kwd>
        <kwd>medical imaging</kwd>
        <kwd>deep learning</kwd>
        <kwd>ImageCLEF benchmarking lab</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>AI systems for medical tasks like predicting, detecting, and classifying diseases rely on the availability of
large and diverse datasets for training. High-quality data allow these models to learn complex patterns
and improve their accuracy and reliability. However, access to real medical data is not easy due to
privacy concerns. Patients are usually willing to share their medical information only for their own
treatment and not for research. This makes it dificult to gather enough data to efectively train AI
models, slowing progress in developing better tools for healthcare care.</p>
      <p>One way to solve this problem is to create synthetic data: artificial data that looks like real medical
data but does not come from actual patients. Generative models, such as Generative Adversarial
Networks (GAN), can be used to create these datasets. Synthetic data can help researchers build and
test AI systems without needing to rely on real patient data, which protects privacy and makes it easier
to collect the variety of information needed for training. But there is an important challenge with
synthetic data: it must not include hidden details, or “fingerprints” from the real data it was trained on.
If synthetic data can somehow be traced back to the original patient data, it could risk exposing private
information. Ensuring that synthetic data are completely free from such “fingerprints” is critical.</p>
      <p>
        This is the third edition of the GANs task, part of ImageCLEF 2025 Lab [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], in which we continue
to investigate the hypothesis that generative models generate synthetic medical images that retain
"fingerprints" from the real images used during their training based on the lessons learned from the
previous two editions:
• In the first [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] and second [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] editions of this task, held at ImageCLEF 2023 [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] and 2024 [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ],
various generative models were analyzed within the framework of the first subtask to investigate
whether synthetic images contained "fingerprints" of the real medical data used during training.
The results demonstrated that the tested generative models do retain and imprint features from
their training data, raising important security and privacy concerns. These findings underscore
the need for robust techniques to detect and mitigate such imprints to ensure that synthetic
images protect patient privacy while maintaining their utility for research and development.
• In the 2nd edition of the task [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], it was confirmed that generative models leave unique
“fingerprints” on the synthetic images they produce. By analyzing images generated from various
models, distinct patterns and features were identified that allowed the attribution of synthetic
images to their respective generative models.
      </p>
      <p>To investigate the potential privacy risks associated with synthetic medical images, the 2025 edition
of the GANs Task focuses on two complementary subtasks. The first subtask challenges participants
to determine whether specific real images were used during the training of a GAN to produce given
synthetic outputs, thereby addressing the risk of image-level information leakage. The second subtask
extends this inquiry by asking participants to attribute synthetic images - generated using a Difusion
model - to one of several predefined subsets of real training data. This focuses on whether broader,
dataset-level characteristics are retained during the generative process. An overview of these two
subtasks is illustrated in Figure 1.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Tasks description</title>
      <sec id="sec-2-1">
        <title>2.1. Sub-task 1: Detect Training Data Usage</title>
        <sec id="sec-2-1-1">
          <title>2.1.1. Description</title>
          <p>
            We continued to investigate the hypothesis that generative models are generating medical images that
are in some way similar to the ones used for training. The task addresses the security and privacy
concerns related to personal medical image data in the context of generating and using artificial images
in diferent real-life scenarios. This edition continues the task proposed in both previous editions [
            <xref ref-type="bibr" rid="ref2 ref3">2, 3</xref>
            ]
by investigating another GAN. The objective of the task is to detect “fingerprints” within the synthetic
biomedical image data to determine which real images were used in training to produce the generated
images. The task consisted in performing analysis of test image datasets and assess which images of
real patients were used for training image generators and which were not. The task is formulated as
follows:
• In this subtask, participants will analyze synthetic biomedical images to determine whether specific
real images were used in the training process of generative models. For each real image in the test set,
participants must label it as either used (1) or not used (0) for generating the given synthetic images.
          </p>
          <p>This task focuses on detecting the presence of training data "fingerprints" within synthetic outputs.</p>
        </sec>
        <sec id="sec-2-1-2">
          <title>2.1.2. Data description</title>
          <p>The benchmarking dataset includes both real and synthetic biomedical images. The real images consist
of axial slices of 3D thoracic CT scans from approximately 8,000 lung tuberculosis patients. These
slices vary in appearance: some may look relatively “normal”, while others exhibit distinct lung lesions,
including severe cases. The real images were provided in 8-bit per pixel PNG format, with dimensions
of 256x256 pixels, providing a standardized resolution for analysis. The synthetic images, also sized
at 256 × 256 pixels, have been generated using a GAN. By providing both real and synthetic datasets,
this task enables participants to analyze and compare the characteristics of synthetic images with their
real counterparts, investigating potential “fingerprints” and patterns related to the training process.
Examples of real and generated images are depicted in Figure 2.</p>
          <p>Train dataset consists of 3 folders:
• “generated” - contains 5,000 synthetic images generated using a GAN.
• “real_used” - contains 100 real images that were used to train the GAN to
produce the synthetic images in the "generated" folder.</p>
          <p>• “real_not_used”- contains 100 real images that were not used for training.
Test dataset was organized as follows:
• “generated”- contains 2,000 additional synthetic images. These images
were generated using the same model trained under the same conditions as
those used to create the synthetic images in the training dataset.
• “real_unknown” - contains a mix of 500 real images. Some of these images
were used in training the generative model, while others were not.</p>
        </sec>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Sub-task 2: Identify Training Data Subsets</title>
        <sec id="sec-2-2-1">
          <title>2.2.1. Description</title>
          <p>In this subtask, participants will link each synthetic biomedical image to the specific subset of real data
used during its generation. The goal is to identify the particular dataset of real images that contributed
to the training of the generative model responsible for creating each synthetic image. This requires
a more detailed attribution of synthetic images to their corresponding training subsets. The task is
formulated as follows:
• In this subtask, participants will link each synthetic biomedical image to the specific subset of real
data used during its generation. The goal is to identify the particular dataset of real images that
contributed to the training of the generative model responsible for creating each synthetic image.</p>
          <p>This requires a more detailed attribution of synthetic images to their corresponding training subsets.</p>
        </sec>
        <sec id="sec-2-2-2">
          <title>2.2.2. Data description</title>
          <p>The benchmarking dataset for this task is similar to the dataset provided for the first subtask in terms
of cohort, acquisition conditions, image dimensions (256× 256) and bit-depth (8 bit per pixel); the only
diference lies in the scan types. In addition to thoracic CTs, it also contains cervical and abdominal CTs,
as exemplified in Figure 3. The synthetic images, also sized at 256 × 256 pixels, have been generated
using a Difusion model. By providing both real and synthetic datasets, this task enables participants to
analyze and compare the characteristics of synthetic images with their real counterparts, investigating
potential “fingerprints” and patterns related to the training process. Examples of generated images are
shown in Figure 3. The training dataset consists of two main folders:
• “generated”- contains 5 subfolders of synthetic images. Each subset was
generated using a different training dataset for the generative model.
• “real” - contains 5 subfolders, each corresponding to a specific training
dataset used to train the generative model. The real images in each
subfolder were used to generate the synthetic images in the corresponding
"generated" subfolder.</p>
          <p>The mapping between the real and generated images is as follows:
Folder “t1” (real images) - Used to generate synthetic images in “gen_t1”
Folder “t2” (real images) - Used to generate synthetic images in “gen_t2”
Folder “t3” (real images) - Used to generate synthetic images in "gen_t3”
Folder “t4” (real images) - Used to generate synthetic images in “gen_t4”
Folder “t5” (real images) - Used to generate synthetic images in “gen_t5”.</p>
          <p>The test dataset contains 25,000 generated images, each derived from a real subgroup of images in the
training dataset. Each image will be assigned a label consistent with those used in the training dataset.</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Evaluation methodology</title>
      <sec id="sec-3-1">
        <title>3.1. Sub-task 1: Detect Training Data Usage</title>
        <p>
          The oficial evaluation metric of the task is Cohen’s kappa score. In addition, the F1-score, accuracy,
precision, and recall were computed to asses the task which can be considered a binary-class classification
problem. Cohen’s kappa [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] is a statistic technique that measures inter-annotator agreement on a
classification problem and it is defined as:
,where  is the empirical probability of agreement on the label assigned to any sample (the observed
agreement ratio), and  is estimated using a per-annotator empirical prior over the class labels.
        </p>
        <p>= ( − )/(1 − )
 1 −  =
  =
 =</p>
        <p>+</p>
        <p>+  
  · 
  + 
 =
  =</p>
        <p>+  
  +   +   +</p>
        <p>+  
where   stands for true positive,   for true negative,   for false positive, and   for false
negative.</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Sub-task 2: Identify Training Data Subsets</title>
        <p>The oficial metric of the task is Accuracy. In addition, Precision, Recall, F1 score and Specificity were
computed.
(1)
(2)
(3)
(4)
(5)
(6)</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Results</title>
      <sec id="sec-4-1">
        <title>4.1. Participation stats</title>
        <p>Overall, 41 teams registered for our task. Of these, 14 teams completed the first subtask by submitting
runs, and 4 teams completed the second subtask. In total, 9 teams submitted working notes papers.
Notably, 2 teams participated in both subtasks, including the task organizing team. The continued
interest in the first subtask, now in its third edition, highlights its ability to attract more participating
teams.</p>
        <p>Each participating team was allowed to submit up to 10 runs per task. We received a total of 95
submitted runs: 91 for Subtask 1 and 14 for Subtask 2. Table 1 presents the list of participating teams
and their institutions. The rankings for Subtask 1 are shown in Table 2, and those for Subtask 2 are
presented in Table 3, limited to the teams that described their methods by submitting working notes
papers.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Presented methods</title>
        <sec id="sec-4-2-1">
          <title>4.2.1. Sub-task 1: Detect Training Data Usage</title>
          <p>
            SCOPE VIT Visioneers [
            <xref ref-type="bibr" rid="ref7">7</xref>
            ] team proposed a framework using a Siamese Neural Network (SNN) to
detect whether real medical images were used in training. The architecture consists of twin subnetworks
based on popular backbones (ResNet50, DenseNet161, EficientNetB2, and ViT), enhanced with a
crossattention mechanism to align feature maps of real and synthetic image pairs. These attention-guided
features are fed into an adaptive similarity module that combines absolute diference and dot product
operations, followed by a multilayer perceptron to score the likelihood that an image contributed to
GAN training. Evaluated on a lung CT dataset, the ResNet50 backbone with cross-attention achieved
the best performance, demonstrating high accuracy and generalizability. All results obtained by the
team are shown in Table 2 and consists in the following methods:
          </p>
          <p>
            • Run ID 1160: Siamese network with a ResNet50 backbone augmented by cross-attention
Challengers [
            <xref ref-type="bibr" rid="ref8">8</xref>
            ] explored four diferent methods. The first two models applied an unsupervised
framework combining ResNet50 for deep feature extraction, Principal Component Analysis (PCA)
for dimensionality reduction, and various clustering algorithms (k-means, DBSCAN (A density based
clustering algorithm capable of identifying arbitrarily shaped clusters and detecting outliers), Gaussian
Mixture Model (GMM), and agglomerative clustering) to analyze similarities between real and synthetic
images. The third model was a supervised binary classifier based on ResNet50V2, fine-tuned via transfer
learning and enhanced with a custom classification head, achieving the best overall performance. Finally,
          </p>
          <p>Afiliation
Vellore Institute of Technology</p>
          <p>San Diego VA</p>
          <p>Health Care System
Sri Sivasubramaniya Nadar College</p>
          <p>of Engineering
Sri Sivasubramaniya Nadar College</p>
          <p>of Engineering
Pune Institute of Technology</p>
          <p>Yunnan University
Yunnan University</p>
          <p>Yunnan University
National University of Science and Technology</p>
          <p>POLITEHNICA Bucharest</p>
          <p>Country</p>
          <p>
            India
US
India
India
India
China
China
China
a fourth hybrid approach combined either deep features or Local Binary Patterns (LBP) with clustering
and distance-based classification, determining training data reuse by measuring the distance between
test samples and learned synthetic image centroids. All results obtained by the team are shown in
Table 2 and consists in the following methods:
• Run ID 1778: Method 1 – ResNet50 + PCA + Clustering (KMeans, GMM, Agglomerative, DBSCAN);
• Run ID 1779: Method 2 – ResNet50 + PCA + Clustering;
• Run ID 1811: Method 3 – ResNet50V2 + Binary Classification (Supervised CNN Approach);
• Run ID 1776: Method 4 – Deep Feature Based Clustering and Distance Based Classification.
Team Medhastra [
            <xref ref-type="bibr" rid="ref9">9</xref>
            ] explored a deep feature similarity-based approach. Their method leverages a
ResNet-50 model, pre-trained on ImageNet [
            <xref ref-type="bibr" rid="ref10">10</xref>
            ], to extract 2048-dimensional embeddings from both real
and generated images. Cosine similarity scores are computed between each synthetic image and the
real image pool, and a statistical threshold (based on mean and standard deviation) is applied to classify
wheater a real image was used for training. Additionaly, a logistic regression model using similarity
features was trained to support this classification. The team submitted one run (Run ID 1288) to the
task, available in Table 2.
          </p>
          <p>
            Neural Nexus [
            <xref ref-type="bibr" rid="ref11">11</xref>
            ] adopted a multi-faceted deep learning approach. Their strategy combined supervised
and unsupervised techniques. They trained multiple autoencoders, including convolutional and Vision
Transformer (ViT)-based variants, to learn latent representations from synthetic images. The ViT
autoencoder, pretrained on a tuberculosis dataset, was used to extract features for clustering and
classification tasks. These latent vectors were clustered using spectral clustering to identify patterns
correlating with the use of specific real images in GAN training. In a parallel track, the team developed
a ResNet-based autoencoder and supplemented the learned features with handcrafted descriptors, such
as GLCM texture features, wavelet transforms, and Gabor filters. These were then input to a Random
Forest classifier that also integrated Mahalanobis distance-based anomaly scores to identify potential
training data influence. Finally, they explored a Dual-Contrastive Learning GAN (DCLGAN) to learn
and compare attention-based feature maps from real and generated images. The overlap in discriminator
attention regions was analyzed to infer training image influence, forming another signal for attribution.
Each component of the pipeline was designed to capture complementary aspects of training data
ifngerprints. Results obtained by the team are shown in Table 2 and consists in the following methods:
• Run ID 1878: ViT-Based Encoder
• Run ID 1880: Spectral Clustering on AutoEncoder Features
• Run ID 1881: ResNet Autoencoder and Feature-Based Detection Framework
Team zhouyijiang1 [
            <xref ref-type="bibr" rid="ref12">12</xref>
            ] explored a ViT-based two-stage detection architecture to identify potential
ifngerprint leakage from training data in GAN-generated medical images. Their system integrates a
coarse-to-fine strategy: in the first stage, a feature approximation module using the FAISS L2 engine
and cosine similarity identifies the top-50 most similar real images for each synthetic sample. Spectral
clustering further refines these candidates to reduce outliers. In the second stage, a Hybrid Contrastive
Attention Network (CANet) performs deep semantic matching. The model fuses features from multiple
ViT layers (6, 12, 18) using a gated fusion mechanism with dynamic weights to generate a
1536dimensional feature embedding. This is passed into a supervised contrastive module combining
crossentropy and contrastive loss to enhance discriminative power and suppress false positives. Feature
encoding is enriched through dynamic data augmentation (including random masking and Gabor noise),
and a custom positional encoding strategy augments anatomical awareness. The deep attention module
consists of a three-stage pipeline: feature mapping, global dependency modeling via an 8-head attention
mechanism, and decision refinement using adaptive temperature scaling and sigmoid classification. A
ifnal decision is made by checking if a sample has high attention-weighted similarity and confidence
&gt;0.6. The method is implemented using ViT-Base and optimized via AdamW with cosine learning rate
decay. The final system balances high detection performance with low false positive rates and real-time
inference. The obtained results are shown in Table 2 and consists in the following:
Team taozi [
            <xref ref-type="bibr" rid="ref13">13</xref>
            ] proposed a hybrid framework. Their approach centered on contrastive learning
and a Mixture of Experts (MoE) architecture that integrates three pre-trained backbones:
ResNet50, EficientNet-B0, and ViT-B/16. Initially, feature similarity matrices were computed using these
backbones to create pools of candidate positives and negatives. These were used to train a MoCo v2-style
contrastive model enhanced with dynamic projection heads and adaptive sample-wise temperature
scaling. The team used an online encoder with momentum updates, InfoNCE loss, and FIFO queues for
negative sampling to ensure diverse and stable contrastive training. Their dynamic projection head
projects 2048-dimensional features into a 256-dimensional embedding space, while generating adaptive
temperature values to control learning sensitivity. A Hybrid Contrastive Loss was employed, balancing
online and momentum components. For multi-model integration, the team implemented a Mixture
of Experts module, which L2-normalized and concatenated outputs from all three backbones. A
twolayer gating network then weighted each exper’s contribution, followed by an 8-head cross-attention
mechanism to refine fused features. This enabled the system to leverage complementary strengths of
each backbone for better generalization across anatomical patterns. The obtained results are shown in
Table 2 and consists in the following:
• Run ID 1875: ResNet50
• Run ID 1874: MoE
• Run ID 1169: EficientNEt-B0
• Run ID 1140: Vision Transformer
• Run ID 1179: ResNet50
• Run ID 1107: ResNet50
Team ZOQ [14] explored a similarity classification approach that combines image enhancement
and deep learning models. The pipeline begins with image preprocessing using four enhancement
techniques: Gaussian filtering to reduce noise while preserving edges, the Laplacian operator for edge
detection, the Hessian matrix to enhance curvature-based features, and bilateral filtering for
edgepreserving smoothing. These transformations aim to highlight critical structural features and improve
the quality of the extracted representations. Following enhancement, the team employed two deep
learning architectures: CNN and ResNet50 for feature extraction. These models were used to generate
high-dimensional feature vectors representing each image. To determine similarity between synthetic
and real images, three diferent similarity metrics were evaluated: Cosine similarity (to assess directional
closeness of feature vectors), Structural Similarity Index (SSIM) (capturing luminance, contrast, and
structural alignment), and Jaccard similarity (measuring overlap between binarized feature regions).
The final classification of an image as "used" or "not used" was determined by applying thresholds to
these similarity scores, ofering a multi-angle view of feature resemblance between real and generated
samples. The obtained results are shown in Table 2 and consists in the following:
• Run ID 1355: CNN + SSIM
• Run ID 1427: ResNet with Jaccard
AI Multimedia Lab [15] explored a twin-branch Siamese network, where each branch received image
inputs and produced feature embeddings via convolutional layers followed by dense layers with L2
normalization. The model was trained using contrastive loss to encourage similar embeddings for
real-synthetic pairs originating from the same training source and dissimilar ones otherwise. Input
pairs were composed of synthetic images and either real-used (positive) or real-not-used (negative)
images. For inference, similarity scores between image pairs were computed using Euclidean distance,
and binary predictions were derived based on learned thresholds. This architecture, previously applied
in Subtask 2 for subset attribution, was adapted here to learn the nuanced similarities indicative of
GAN training data usage. The obtained results are presented in Table 2 under Run ID 1696 and 1492.
          </p>
        </sec>
        <sec id="sec-4-2-2">
          <title>4.2.2. Sub-task 2: Identify Training Data Subsets</title>
          <p>
            SDVAHCS/UCSD [16] presented an ensemble-based deep learning approach to identify the origins
of synthetic biomedical images by classifying them according to the specific real data subsets used in
their generation. The authors employed multiple EficientNet architectures (b0, b1, b2), leveraging a
six-class classification setup to distinguish between real images and five sets of synthetic images, each
linked to a distinct training subset. A max-voting ensemble strategy was used to enhance robustness,
and pseudo-labeling was incorporated to utilize unlabeled test data by iteratively assigning labels and
retraining models. In one configuration, a portion of the original data was sequestered to validate the
generalization of the pseudo-labeling technique and to detect overfitting. The results obtained by the
team are shown in Table 3 and consists in the following methods:
• Run ID 1425: A single EficientNet-b1 model trained with a learning rate of 0.0001 and batch size
32; selected for best validation performance.
Team Medhastra [17] employed a ResNet-18-based deep learning pipeline to classify. The approach
involved two main components: feature extraction and supervised multi-class classification. For feature
extraction, ResNet-18 (pretrained on ImageNet [
            <xref ref-type="bibr" rid="ref10">10</xref>
            ]) was repurposed by removing its classification
layer, transforming each image into a 512-dimensional vector capturing high-level semantic features.
These embeddings were used to measure similarity between synthetic and real images and support
subgroup inference. For classification, the team fine-tuned ResNet-18 to perform five-way classification.
The model was trained on labeled real images and validated using synthetic images to better simulate
inference-time conditions. During inference, the trained model predicted subgroup labels for unlabeled
synthetic images in the test set. The results of the team are available Table 2 under Run ID 1287.
AI Multimedia Lab [15] employed two distinct methods: (i) Method 1 –Feature Clustering, their
previous feature-based clustering pipeline [18] was adaptedto the 2025 subset attribution task. The
approach begins with feature extraction using four pretrained convolutional models: MobileNetV2,
ResNet50, EficientNet, and DenseNet, all originally trained on ImageNet [
            <xref ref-type="bibr" rid="ref10">10</xref>
            ]. These models generate
high-dimensional embeddings from the input images to capture their semantic content. To infer
which subset (T1-T5) was used to generate each synthetic image, the team experimented with three
classification strategies: k-means clustering, hierarchical clustering, and Support Vector Machines. The
ifrst two are unsupervised techniques that group similar images based on feature similarity - k-means via
centroid minimization and hierarchical clustering via distance-based tree structures while SVM serves
as a supervised baseline trained on labeled subset data. The models were fine-tuned on the training
data, with 10% of the set held out for validation to evaluate clustering and classification performance.
This method investigates whether deep features from pretrained networks can efectively capture
patterns indicative of the GAN training subsets, using both supervised and unsupervised grouping
strategies. ii) A Siamese network was trained using contrastive loss to learn whether a real-synthetic
image pair came from the same training subset. Each synthetic image was compared with real images
from all five subsets, and average embedding distances were calculated per subset. The subset with the
smallest average distance was selected as the predicted origin. This approach focuses on learning visual
similarity through a relational, pairwise metric. The results of the team are available in Table 3:
• Run ID 1396: Siamese Neural Network
• Run ID 1271: Densenet + SVM
• Run ID 1269: EficientNet + SVM
• Run ID 1268: ResNet + SVM
• Run ID 1267: MobileNetV2 + SVM.
          </p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Discussion</title>
      <p>A wide range of strategies was deployed for Subtask 1, spanning from contrastive learning and Siamese
architectures to hybrid clustering approaches, vision transformers, and handcrafted feature modeling.</p>
      <p>SCOPE VIT Visioneers adopted a SNN enhanced with a cross-attention module and an adaptive
similarity head, using backbones such as ResNet50 and ViT. Despite the architectural sophistication, their
best-performing submission (Run ID 1160) achieved a Cohen’s kappa of -0.032, indicating performance
worse than chance, despite achieving a moderate accuracy of 0.484. Similarly, Medhastra employed
cosine similarity thresholds and logistic regression on ResNet50 embeddings, but their submission
(Run ID 1288) also yielded a negative kappa score of 0.016, further confirming the dificulty of the
task. AI Multimedia Lab, which previously applied a Siamese approach to Subtask 2, adapted the same
architecture here, but their runs (IDs 1696 and 1492) achieved low kappa scores of 0.036 and -0.044,
respectively. These results highlight that feature-level similarity, even when trained with contrastive
loss, may be insuficient to reliably detect GAN training inclusion at the image level. Challengers
investigated both unsupervised and supervised pipelines, including PCA-reduced clustering and a
ifne-tuned ResNet50V2 classifier. While their best supervised model (Run ID 1811) performed relatively
well in terms of accuracy (0.506), it still produced a very low kappa score of 0.012, suggesting that even
supervised classifiers struggle with consistent predictions beyond chance. Neural Nexus introduced
an ambitious multi-track pipeline combining ViT-based autoencoders, handcrafted features, and GAN
attention analysis. Their best run (ID 1878) achieved the highest Cohen’s kappa of 0.148, still modest
but indicating some ability to detect data usage patterns, supported by a higher F1-score of 0.5864. Team
zhouyijiang1 implemented a ViT-based architecture featuring coarse-to-fine filtering, spectral clustering,
and a hybrid contrastive attention network. Their suite of runs (IDs 1801-1804, 1873) consistently
achieved kappa values in the 0.128-0.136 range, among the best overall, demonstrating slight but stable
agreement beyond chance. Taozi pursued a contrastive learning framework enriched with a Mixture of
Experts (MoE) module integrating ResNet, EficientNet, and ViT backbones. Despite their architectural
complexity, none of their runs exceeded a kappa of 0.108, and several dropped below zero, reflecting
the dificulty of leveraging contrastive embeddings for precise image-level attribution. ZOQ applied
deep similarity classification using image enhancement (e.g., Gaussian, Laplacian, Hessian filtering) and
various similarity metrics (cosine, SSIM, Jaccard). Their submissions, including Run IDs 1355 and 1427,
achieved Cohen’s kappa values ranging from -0.016 to -0.132, reinforcing that even enriched structural
representations could not capture consistent training reuse signals.</p>
      <p>Cohen’s kappa scores across all submissions were notably low, with only a few teams surpassing the
0.1 threshold. This statistical measure, which ranges from -1 (systematic disagreement) to 1 (perfect
agreement), reflects the agreement between predicted labels and ground truth while correcting for
chance. The prevalence of negative or near-zero kappa values indicates that most models performed at
or below chance level, implying methods’ inability to detect the source images used for training. This
also could reflect model’s ability to generate synthetic images that do not contain “fingerprints” that
can be traced back to the source of training.</p>
      <p>In Subtask 2, participants were tasked with linking each synthetic biomedical image to the specific
subset of real data used during its generation. This required detailed attribution of synthetic images
to one of five real training subsets (T1-T5), posing a significant challenge in capturing subtle
datasetspecific features embedded by the generative model. The top performance, excluding the organizing
team was obtained by SDVAHCS/UCSD team (agentili), submitting multiple ensemble-based runs. Their
most notable configurations (Run IDs 1782, 1871, and 1883) each achieved an accuracy of 98.8%. These
submissions relied on ensembles of EficientNet models (b0 - b2) and employed a max-voting strategy
for robust decision-making. Pseudo-labeling was introduced to incorporate unlabeled test data by
iteratively assigning labels and retraining the models. Notably, Run ID 1782 included a sequestered
subset of the training data to monitor and prevent overfitting during pseudo-label-based augmentation.
This strategy enabled the team to verify that the pseudo-labeling process did not lead to data leakage or
artificial performance inflation, enhancing model generalization and credibility. A simpler configuration
(Run ID 1425) using a single EficientNet-b1 model also performed strongly with 97.08% accuracy.</p>
      <p>The organizing team, AI Multimedia Lab, whose submission (Run ID 1396) achieved the highest
accuracy of 99.04% was based on a SNN trained with contrastive loss to learn similarity relationships
between synthetic and real images. During inference, the model compared each synthetic image to
examples from each subset and attributed the image to the subset with the smallest average embedding
distance. This pairwise metric-learning approach efectively captured fine-grained inter-image
similarities, yielding highly discriminative performance across all five subsets. Team Medhastra (Run ID
1287) adopted a more streamlined approach using a ResNet-18 backbone. Their pipeline combined deep
feature extraction and supervised multi-class classification. Feature embeddings were generated from
ResNet-18, and the model was fine-tuned to distinguish between the five subsets using real training
images. During inference, predictions for synthetic images were generated directly via softmax
classification. Despite its relative simplicity and lack of ensembling, this method achieved a commendable
94.84% accuracy, demonstrating the eficacy of targeted fine-tuning on well-designed architectures.</p>
      <p>In contrast, AI Multimedia Lab’s second method, which extended their prior year’s pipeline [18],
yielded lower performance. This approach (Run IDs 1267 - 1271) relied on deep feature clustering
and classification using models such as MobileNetV2, ResNet50, EficientNet, and DenseNet. The
extracted embeddings were processed via various clustering or classification techniques, including
k-means, hierarchical clustering, and SVMs. While methodologically diverse and interpretable, these
models achieved accuracies ranging from 41.12% to 52.36%, indicating that static pretrained features
lacked the discriminative capacity required for this fine-grained attribution task. The relatively poor
performance suggests that such representations may be insuficient for capturing the generative nuances
that distinguish training subsets in GAN outputs.</p>
      <p>Overall, the results indicate that contrastive learning frameworks and ensemble deep learning
models, particularly those using EficientNet architectures, are highly efective for subset attribution.
Techniques such as pseudo-labeling, max-voting, and overfitting control through sequestered validation
further enhance performance. Meanwhile, feature clustering methods, though appealing for their
simplicity and generality, may require significant enhancement or task-specific tuning to match the
performance of supervised or metric-learning-based approaches.</p>
      <p>Lessons learned</p>
      <p>The 2025 edition of the ImageCLEFmedical GANs task provided valuable insights into the challenges
of detecting training data exposure in synthetic medical images, with each subtask ofering lessons
from a diferent generative modeling approach.</p>
      <p>Subtask 1 focused on detecting whether individual real images had been used to train GAN. Despite
the use of a wide array of methods – including Siamese networks, contrastive learning, and
attentionbased models –participants achieved only modest performance, with the highest Cohen’s kappa score
reaching 0.148. This suggests that detecting image-level membership in GAN training data remains a
highly challenging problem. It may also indicate that, under the given task conditions, the GAN was
efective at generating images that do not exhibit easily detectable traces of specific training samples.
This is a potentially encouraging result from a privacy standpoint, but it simultaneously underscores the
limitations of current detection frameworks and the need for continued exploration of more sensitive
or robust analysis techniques.</p>
      <p>Subtask 2, in contrast, explored attribution at a broader scale: identifying the training subset used
to generate synthetic images from a difusion model. Here, performance was significantly higher,
with multiple teams exceeding 98% accuracy. While this task did not involve image-level membership
inference, it revealed that difusion-generated images can still reflect statistical characteristics of the
training data subsets. The success of various supervised and semi-supervised classification techniques
suggests that difusion models, while efective at preserving image realism, may still encode latent
information about their source data distributions. This raises important considerations about the use
of synthetic data for data sharing or augmentation, particularly when the provenance or diversity of
training data must remain confidential.</p>
      <p>Collectively, the lessons from both subtasks emphasize that generative model outputs can leak
diferent types of information, depending on both the model architecture and the nature of the detection task.
Continued benchmarking and task diversification will be essential for building a more comprehensive
understanding of privacy risks in generative medical imaging and for guiding the development of
responsible and secure data generation practices.</p>
    </sec>
    <sec id="sec-6">
      <title>6. Conclusions</title>
      <p>The 2025 edition of the ImageCLEFmedical GANs Task further investigated the privacy and security
implications of using GAN-generated synthetic medical images by addressing two key challenges:
determining whether specific real images were used during GAN training (Subtask 1), and attributing
synthetic images to their original training data subsets (Subtask 2).</p>
      <p>In Subtask 1, despite the broad range of approaches: from SNN and contrastive learning to
attentionbased ViTs and handcrafted similarity metrics, overall performance remained low. The highest Cohen’s
kappa score reached only 0.148, suggesting that current methodologies struggle to reliably detect
whether individual real images were used to generate synthetic outputs. These results may indicate the
capability of the employed generative model to produce realistic images that do not easily leak training
data, thereby better preserving patient privacy. However, the challenge also highlights a pressing
need for more refined evaluation techniques and robust detection frameworks capable of identifying
subtle "fingerprints" in high-fidelity synthetic medical imagery. This remains an open area for further
investigation.</p>
      <p>In contrast, Subtask 2 achieved considerably better results, with multiple teams surpassing 98%
classification accuracy. The most successful approaches utilized supervised classification with EficientNet
ensembles and contrastive Siamese networks, some enhanced with pseudo-labeling and max-voting
strategies. These findings suggest that while identifying whether an individual image was used in GAN
training remains dificult, it is more feasible to capture coarse-grained, dataset-level attributions
embedded in synthetic images generated by Difusion models. This supports the hypothesis that Difusion
models may retain latent features reflecting the broader statistical properties of their training subsets.</p>
      <p>Overall, the task underscores the need for ongoing research into privacy-preserving generative
modeling, especially in sensitive domains such as healthcare. Benchmarking challenges like this remain
crucial for understanding the evolving capabilities and risks associated with synthetic medical data. We
will continue investigating these issues in future editions of ImageCLEF, where we plan to introduce
new tasks aimed at further exploring generative model privacy, attribution, and generalization across
modalities and model types.</p>
    </sec>
    <sec id="sec-7">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the authors used ChatGPT-4o in order to: Grammar and spelling
check and improve writing style. After using this tool, the authors reviewed and edited the content as
needed and take full responsibility for the publication’s content.</p>
    </sec>
    <sec id="sec-8">
      <title>7. Acknowledgments</title>
      <p>The work of Alexandra Andrei was supported by a grant of the Ministry of Research, Innovation and
Digitization, CCCDI - UEFISCDI, project number PN-IV-P6-6.3-SOL-2024-0049, within PNCDI IV. The
work of Mihai Gabriel Constantin was supported by a grant of the Ministry of Research, Innovation
and Digitization, CCCDI - UEFISCDI, project number PN-IV-P6-6.3-SOL-2024-0060, within PNCDI IV.
[14] H. Zuo, X. Zhou, Evaluating of the privacy of images generated by imageclefmedical gan 2025
using similarity classification method based on image enhancement and deep learning mode, in:
CLEF2025 Working Notes, Madrid, Spain, 2025.
[15] A. Andrei, M. G. Constantin, M. Dogariu, L. Ştefan, B. Ionescu, Ai multimedia lab at
imageclefmedical gans 2025: Identifying real-image usage in generated medical images, in: CLEF2025 Working
Notes, Madrid, Spain, 2025.
[16] A. Gentili, Identifying the origins of synthetic biomedical images: Anensemble approach with
pseudo-labeling, in: CLEF2025 Working Notes, Madrid, Spain, 2025.
[17] S. Chandrasekar, V. R. S, V. P, Identify training data subsets in gan-generated medical images, in:</p>
      <p>CLEF2025 Working Notes, Madrid, Spain, 2025.
[18] A. Andrei, M. G. Constantin, M. Dogariu, B. Ionescu, Ai multimedia lab at imageclefmedical gans
2024: Deep learning approaches for analyzing synthetic medical images, in: CLEF2024 Working
Notes,CEUR Workshop Proceedings, Grenoble, France, 2024.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>B.</given-names>
            <surname>Ionescu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Müller</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.-C.</given-names>
            <surname>Stanciu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.-G.</given-names>
            <surname>Andrei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Radzhabov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Prokopchuk</surname>
          </string-name>
          , Ştefan, LiviuDaniel, M.-G. Constantin,
          <string-name>
            <given-names>M.</given-names>
            <surname>Dogariu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Kovalev</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Damm</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Rückert</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. Ben</given-names>
            <surname>Abacha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>García Seco de Herrera</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. M.</given-names>
            <surname>Friedrich</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Bloch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Brüngel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Idrissi-Yaghir</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Schäfer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. S.</given-names>
            <surname>Schmidt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T. M. G.</given-names>
            <surname>Pakull</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Bracke</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Pelka</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Eryilmaz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Becker</surname>
          </string-name>
          , W.-W. Yim,
          <string-name>
            <given-names>N.</given-names>
            <surname>Codella</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. A.</given-names>
            <surname>Novoa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Malvehy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Dimitrov</surname>
          </string-name>
          ,
          <string-name>
            <surname>R. J. Das</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          <string-name>
            <surname>Xie</surname>
            ,
            <given-names>H. M.</given-names>
          </string-name>
          <string-name>
            <surname>Shan</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Nakov</surname>
            , I. Koychev,
            <given-names>S. A.</given-names>
          </string-name>
          <string-name>
            <surname>Hicks</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Gautam</surname>
            ,
            <given-names>M. A.</given-names>
          </string-name>
          <string-name>
            <surname>Riegler</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          <string-name>
            <surname>Thambawita</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Halvorsen</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Fabre</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Macaire</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Lecouteux</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Schwab</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Potthast</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Heinrich</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Kiesel</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Wolter</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Stein</surname>
          </string-name>
          , Overview of imageclef 2025:
          <article-title>Multimedia retrieval in medical, social media and content recommendation applications, in: Experimental IR Meets Multilinguality</article-title>
          , Multimodality, and
          <string-name>
            <surname>Interaction</surname>
          </string-name>
          ,
          <source>Proceedings of the 16th International Conference of the CLEF Association (CLEF</source>
          <year>2025</year>
          ), Springer Lecture Notes in Computer Science LNCS, Madrid, Spain,
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>A.-G.</given-names>
            <surname>Andrei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Radzhabov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I.</given-names>
            <surname>Coman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Kovalev</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Ionescu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Müller</surname>
          </string-name>
          ,
          <article-title>Overview of imageclefmedical gans 2023 task: identifying training data “fingerprints” in synthetic biomedical images generated by gans for medical image security</article-title>
          ,
          <source>in: Working Notes of the Conference and Labs of the Evaluation Forum (CLEF</source>
          <year>2023</year>
          ), volume
          <volume>3497</volume>
          ,
          <year>2023</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>A.</given-names>
            <surname>Andrei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Radzhabov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Karpenka</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Prokopchuk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Kovalev</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Ionescu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Müller</surname>
          </string-name>
          ,
          <article-title>Overview of 2024 imageclefmedical gans task-investigating generative models' impact on biomedical synthetic images</article-title>
          ,
          <source>in: CLEF2024 Working Notes, CEUR Workshop Proceedings</source>
          , CEUR-WS. org, Grenoble, France,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>B.</given-names>
            <surname>Ionescu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Müller</surname>
          </string-name>
          ,
          <string-name>
            <surname>A.-M. Drăgulinescu</surname>
            ,
            <given-names>W.-W.</given-names>
          </string-name>
          <string-name>
            <surname>Yim</surname>
            ,
            <given-names>A. Ben</given-names>
          </string-name>
          <string-name>
            <surname>Abacha</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Snider</surname>
            , G. Adams,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Yetisgen</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Rückert</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>García Seco de Herrera</surname>
          </string-name>
          , et al.,
          <article-title>Overview of the imageclef 2023: multimedia retrieval in medical, social media and internet applications</article-title>
          ,
          <source>in: International Conference of the Cross-Language Evaluation Forum for European Languages</source>
          , Springer,
          <year>2023</year>
          , pp.
          <fpage>370</fpage>
          -
          <lpage>396</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>B.</given-names>
            <surname>Ionescu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Müller</surname>
          </string-name>
          ,
          <string-name>
            <surname>A.-M. Drăgulinescu</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Rückert</surname>
            ,
            <given-names>A. Ben</given-names>
          </string-name>
          <string-name>
            <surname>Abacha</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>García Seco de Herrera</surname>
            , L. Bloch,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Brüngel</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Idrissi-Yaghir</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Schäfer</surname>
          </string-name>
          , et al.,
          <source>Overview of the imageclef</source>
          <year>2024</year>
          :
          <article-title>Multimedia retrieval in medical applications</article-title>
          ,
          <source>in: International Conference of the Cross-Language Evaluation Forum for European Languages</source>
          , Springer,
          <year>2024</year>
          , pp.
          <fpage>140</fpage>
          -
          <lpage>164</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>J.</given-names>
            <surname>Cohen</surname>
          </string-name>
          ,
          <article-title>A coeficient of agreement for nominal scales</article-title>
          ,
          <source>Educational and psychological measurement 20</source>
          (
          <year>1960</year>
          )
          <fpage>37</fpage>
          -
          <lpage>46</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>F.</given-names>
            <surname>Areeb</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Vashist</surname>
          </string-name>
          , L. Kalinathan,
          <article-title>Controlling the quality of synthetic medical images created via gans</article-title>
          ,
          <source>in: CLEF2025 Working Notes</source>
          , Madrid, Spain,
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>K. S. S</given-names>
            ,
            <surname>M. S</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Narayanan</surname>
          </string-name>
          ,
          <string-name>
            <surname>B. P</surname>
          </string-name>
          , K. Ayyamperumal,
          <article-title>Detecting training data usage in synthetic images using machine learning techniques</article-title>
          ,
          <source>in: CLEF2025 Working Notes</source>
          , Madrid, Spain,
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>S.</given-names>
            <surname>Chandrasekar</surname>
          </string-name>
          ,
          <string-name>
            <surname>V. R. S</surname>
          </string-name>
          , V. P,
          <article-title>Detecting training data fingerprints in gan-generated medical images</article-title>
          ,
          <source>in: CLEF2025 Working Notes</source>
          , Madrid, Spain,
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>J.</given-names>
            <surname>Deng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Dong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Socher</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.-J.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Fei-Fei</surname>
          </string-name>
          ,
          <article-title>Imagenet: A large-scale hierarchical image database</article-title>
          ,
          <source>in: 2009 IEEE Conference on Computer Vision and Pattern Recognition</source>
          ,
          <year>2009</year>
          , pp.
          <fpage>248</fpage>
          -
          <lpage>255</lpage>
          . doi:
          <volume>10</volume>
          .1109/CVPR.
          <year>2009</year>
          .
          <volume>5206848</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>M.</given-names>
            <surname>Barve</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Nambiar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Bhedasgaonkar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Date</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I.</given-names>
            <surname>Shah</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. G. V.</given-names>
            <surname>Kale</surname>
          </string-name>
          ,
          <article-title>Reverse engineering generative fingerprints in medical images: A deep learning approach to training data attribution</article-title>
          ,
          <source>in: CLEF2025 Working Notes</source>
          , Madrid, Spain,
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Ding</surname>
          </string-name>
          ,
          <article-title>Vit-based generative model fingerprinting</article-title>
          ,
          <source>in: CLEF2025 Working Notes</source>
          , Madrid, Spain,
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>D.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <article-title>Evaluation of the privacy of images generated by imageclefmedical gans 2025 based on pre-trained model feature extraction methods</article-title>
          ,
          <source>in: CLEF2025 Working Notes</source>
          , Madrid, Spain,
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>