<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Comparative Evaluation of StyleGAN3-Based Augmentation Strategies for Enhanced Medical Image Classification</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Faycal Touazi</string-name>
          <email>f.touazi@univ-boumerdes.dz</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Djamel Gaceb</string-name>
          <email>d.gaceb@univ-boumerdes.dz</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Amira Tadrist</string-name>
          <email>a.tadrist@univ-boumerdes.dz</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Sara Bakiri</string-name>
          <email>s.bakiri@univ-boumerdes.dz</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>LIMOSE Laboratory, Computer Science Department, University M'hamed Bougara</institution>
          ,
          <addr-line>Independence Avenue, 35000 Boumerdes</addr-line>
          ,
          <country country="DZ">Algeria</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Deep learning models for medical image classification face significant challenges due to class imbalance and the limited availability of annotated datasets, particularly for rare diseases. Traditional data augmentation techniques, such as rotation, translation, etc., often fail to provide sufficient diversity to perform a good classification for minor classes. To address this issue, various strategies have been explored, including oversampling, undersampling, cost-sensitive learning, and synthetic data generation using generative adversarial networks (GANs). In this study, we evaluate the impact of using a generative AI based approaches and demonstrate that the most effective strategy is to combine synthetic augmentation with traditional methods. Specifically, we employ StyleGAN3 to generate high-fidelity synthetic images that, when integrated with traditional data-augmentation techniques, may improve the performance of deep learning models on medical image classification. We validate our method on datasets, including COVID-19 chest X-rays and HAM10000. Experimental results show that this hybrid approach leads to an improvement in classification accuracy, particularly for minority classes, surpassing standalone augmentation strategies. Our findings highlight the potential of AI-driven synthetic data generation as a complementary solution to traditional augmentation, offering a more balanced and diverse dataset for medical image analysis.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Medical imaging</kwd>
        <kwd>Data augmentation</kwd>
        <kwd>Generative Adversarial Networks</kwd>
        <kwd>StyleGAN</kwd>
        <kwd>Class imbalance 1</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>In the field of deep learning for medical imaging, one of the significant challenges is class imbalance
coupled with small annotated data. It is especially challenging when working with rare diseases,
where the low occurrence and brief duration of the appearance of symptoms can make the collect of
data difficult, which may affect the quality of the model training.</p>
      <p>As a result, such a deficiency affects classification model performance, particularly in classifying
complicated pathologies that are important even their rarity in the datasets. This asymmetry degrades
the performance of the model in favor of the majority classes, thus decreasing the precision and
reliability of the predictions compared to the minority classes.</p>
      <p>The imbalance datasets, coupled with the limited availability of annotated datasets, presents
obstacles to the development of efficient and high-quality models in the field of medical imaging.
Traditional classification models become overfitted to the dominant classes, leading to a significant
loss of accuracy for the minority classes. While traditional data augmentation techniques such as
rotation, resizing, and cropping—are often applied to alleviate this issue, they often fail to generate the
necessary diversity and do not notably enhance the generalization capacity of the models.</p>
      <p>
        Classical data augmentation techniques [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], such as rotation, flipping, scaling, and cropping, are
widely used to artificially increase the size of training datasets and improve model generalization.
These methods help in introducing minor variations to the images, making the model more robust to
small transformations. However, they have significant limitations, especially in the medical imaging
domain. Since medical images often contain complex and subtle patterns that are crucial for diagnosis,
simple transformations may not sufficiently capture the variability needed to enhance model
performance. Furthermore, these techniques do not create new pathological patterns but merely
0000-0001-5949-5421 (F. Touazi); 0000-0002-6178-0608 (D. Gaceb)
© 2025 Copyright for this paper by its authors.
      </p>
      <p>Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0).
modify existing ones, limiting their effectiveness in addressing class imbalance. As a result, they may
not significantly improve the classification of rare diseases, which require more sophisticated
augmentation strategies capable of generating realistic and diverse samples.</p>
      <p>
        To address class imbalance in deep learning, various strategies can be employed (see [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] for an
exhaustive review).
      </p>
      <p>
        1. Oversampling Methods: Oversampling techniques aim to increase the representation of
minority- class samples to balance the dataset. Synthetic Minority Over-sampling Technique
(SMOTE) and its variants generate synthetic data points to improve class distribution [
        <xref ref-type="bibr" rid="ref3 ref4">3, 4</xref>
        ].
Additionally, data augmentation techniques introduce transformed versions of existing images to
improve model generalization. While oversampling has been shown to be one of the most effective
techniques for CNN-based classification [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], its effectiveness can be limited on images in general
and for medical imaging.
2. Undersampling Methods: In contrast, undersampling reduces the number of majority-class
samples to achieve a more balanced dataset, thereby improving class proportions and reducing
computational cost. Random undersampling removes a subset of majority samples, while more
advanced techniques, such as cluster-based undersampling, aim to retain the most informative
samples. Although undersampling is effective in extreme imbalance scenarios, it may lead to
information loss, particularly in complex medical datasets [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ].
3. Other Learning Approaches: Beyond sampling strategies, cost-sensitive learning modifies the
loss function to assign higher penalties for misclassifications in the minority class, with focal loss
being a notable example that prioritizes hard-to-classify instances [
        <xref ref-type="bibr" rid="ref6 ref7">6, 7</xref>
        ]. Ensemble learning
improves prediction accuracy by combining multiple classifiers, but its high computational cost
can be prohibitive [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. Hybrid approaches integrate data-level and algorithm-level solutions, such
as clustering with sampling techniques or cost-sensitive learning with neural networks.
Semisupervised and self-supervised learning leverage unlabeled data to enhance feature representation
and generalization, while deep metric learning and contrastive learning focus on learning more
discriminative representations without altering class distribution. Each of these methods has
trade-offs in efficiency, robustness, and complexity, so a choice can be made based on the nature of
the dataset and the nature of application demands.
      </p>
      <p>
        To overcome the limitations of traditional imbalance-handling techniques, AI-based
imagegeneration methods [
        <xref ref-type="bibr" rid="ref10 ref11 ref9">9, 10, 11, 12, 13</xref>
        ], and particularly StyleGAN [14, 15, 16], have emerged as a
groundbreaking solution for generating realistic synthetic medical images. StyleGAN’s ability to
produce high-fidelity images enables dataset augmentation without compromising the valuable
pathological characteristics essential for medical imaging. By generating samples for
underrepresented classes, StyleGAN helps mitigate class imbalances and improves the stability of
classification models.
      </p>
      <p>In this work, we introduce a StyleGAN3-based data augmentation approach that combines
stateof-the- art generative modeling with classical balancing techniques to enhance the diversity and
representation of minority-class samples. StyleGAN3, with its improved spatial coherence, is
particularly well-suited for medical image synthesis, preserving intricate morphological features of
pathological conditions. Unlike conventional oversampling methods that risk overfitting, our
approach generates diverse and realistic synthetic samples, enriching the dataset and improving the
generalization of classification models. By training StyleGAN3 on the minority class, we aim to
restore class balance, enhance dataset variability, and ultimately improve the robustness of medical
image classification systems.</p>
      <p>This paper is structured as follows: Section 2 is a literature review of data augmentation and GANs
in medical imaging. Section 3 is a description of the methodology of our work, including data
preprocessing, model architecture, and performance metrics. Section 4 is the experimental results and
their discussion. Section 5 concludes the paper and gives future directions of research</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Works</title>
      <sec id="sec-2-1">
        <title>2.1. Based on the Covid_19 Radiography Dataset</title>
        <p>Abdul Waheed et al. [17] introduced CovidGAN, an ACGAN-based model generating synthetic chest
X-ray (CXR) images to address data scarcity in medical imaging. Trained on three datasets (IEEE
Covid Chest X-ray, COVID-19 Radiography Database, COVID-19 Chest X-ray Dataset), CovidGAN
improved CNN classification accuracy from 85% to 95% with augmented data.</p>
        <p>Sharmila V J et al. [18] proposed a DCGAN-CNN hybrid for classifying CXR images (normal,
pneumonia, COVID-19). The DCGAN generates 64×64 synthetic images, later resized for classification. The
CNN, comprising eight convolutional layers, achieved accuracy between 94.8% and 98.6%, surpassing
AlexNet and GoogLeNet.</p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Based on the HAM10000 Dataset</title>
        <p>Bilal Ahmad et al. [19] developed TED-GAN, a hybrid VAE-GAN approach for skin lesion image
generation. Using a dual-GAN framework, their model significantly improved melanoma
classification, increasing sensitivity from 53% to 82% and specificity from 75% to 94%.</p>
        <p>Qinchen Su et al. [20] introduced STGAN, a GAN-based augmentation method for multi-class
imbalanced skin lesion classification. Trained on the HAM10000 dataset, it improved FID, Inception
Score, Precision, and Recall over StyleGAN2 and achieved an accuracy of 98.23% with a ResNet50
classifier.</p>
      </sec>
      <sec id="sec-2-3">
        <title>2.3. Based on Other Datasets</title>
        <p>Bilal Ahmad et al. [21] proposed VAE-GAN, leveraging informative noise instead of Gaussian noise
for brain tumor image generation. Applied to 3,064 CE-MR images, their approach boosted
classification accuracy from 72.63% to 96.25%.</p>
        <p>Guilherme C et al. [22] implemented StyleGAN2-ADA, enhancing image quality for fundus
imaging via adaptive discriminator augmentation to mitigate data scarcity and imbalance.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Proposed Approach</title>
      <sec id="sec-3-1">
        <title>Dataset</title>
        <p>COVID-19 Radiography
COVID-19 Radiography
HAM10000</p>
        <p>HAM10000
CE-MR Brain Tumor
Fundus Imaging</p>
      </sec>
      <sec id="sec-3-2">
        <title>Accuracy</title>
        <p>95%
94.8%-98.6%
Sensitivity: 82%
98.23%
96.25%
85%
Our approach stands out by leveraging GANs not only for data augmentation but also for dataset
balancing. To ensure fair evaluation and prevent data leakage, our dataset was initially divided into
80% for training and 20% for testing. This split remains consistent across all experiments, and
transformations are applied only to the test set. We conducted our study on two medical imaging
datasets: HAM10000 [23] and COVID-19 Radiography [24]. We propose four augmentation strategies:
1. Approach 1 - Traditional Data Augmentation Classical transformations (rotation, flipping,
resizing) are applied to enhance training data diversity.
2. Approach 2 - Targeted Augmentation for Balancing Augmentation is applied specifically to
minority classes to balance the dataset.
3. Approach 3 - StyleGAN3-Based Augmentation with Batch Injection Synthetic images
generated by StyleGAN3 are injected into training batches to improve diversity.
4. Approach 4 - Hybrid StyleGAN3 Augmentation and Traditional Balancing A percentage of
StyleGAN3-generated images is added, followed by traditional balancing techniques.</p>
        <sec id="sec-3-2-1">
          <title>3.1. Approach 1: Traditional Data Augmentation</title>
          <p>We apply standard transformations such as rotation, horizontal/vertical flipping, and color jittering
before feeding images into ResNet50 and InceptionV3. Table 2 summarizes the transformations.</p>
        </sec>
        <sec id="sec-3-2-2">
          <title>3.2. Approach 2: Targeted Augmentation for Balancing</title>
          <p>This approach is inspired by Random Over-Sampling (ROS), a common technique for handling
imbalanced datasets by duplicating samples from minority classes to match the distribution of majority
classes. However, instead of simply duplicating existing images, we apply targeted data augmentation
techniques (e.g., rotation, scaling, contrast adjustments) to generate new synthetic samples. The
gener- ated images are saved and used to balance the dataset, ensuring that minority classes have the
same number of images as the majority classes.</p>
        </sec>
        <sec id="sec-3-2-3">
          <title>3.3. Approach 3: StyleGAN3-Based Augmentation with Batch Injection</title>
          <p>StyleGAN3 is used to generate high-quality synthetic medical images that are directly injected into
training batches during model training. Unlike traditional augmentation, which applies
transformations to existing images, StyleGAN3 synthesizes new samples that mimic the distribution
of real medical images.</p>
          <p>In this approach, synthetic images are generated before training and dynamically included in
minibatches alongside real images. This ensures that the model learns robust representations by exposing
it to a more diverse dataset. The dataset split remains unchanged, with synthetic images used only
during training, preventing bias in the test evaluation.</p>
        </sec>
        <sec id="sec-3-2-4">
          <title>3.4. Approach 4: Hybrid StyleGAN3 Augmentation and Traditional Balancing</title>
          <p>This approach combines StyleGAN3-generated images with traditional dataset balancing techniques
to optimize model performance. The augmentation process consists of two steps:
1. Generation of Synthetic Images: StyleGAN3 is used to generate additional images. We test
four strategies by adding synthetic samples to the original training dataset, increasing the minority
classes by 10%, 20%, 30%, and 40%, respectively.
2. Traditional Balancing Techniques: Once the synthetic images are added, classical balancing
methods are applied. This includes oversampling the minority class and targeted augmentations
(rotation, flipping, and intensity scaling) to equalize class representation.</p>
          <p>This hybrid strategy ensures that the dataset remains well-balanced while introducing new
variations through GAN-generated samples. The classifier is trained on the augmented dataset using
ResNet50 and InceptionV3, and performance is evaluated based on classification metrics.</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Results of data-augmentation</title>
      <sec id="sec-4-1">
        <title>4.1. Results of StyleGAN3</title>
        <p>The generated images show appreciable diversity in the characteristics of each class. This diversity is
crucial to avoid overfitting and to improve the generalization of classification models by including
realistic variations in the data.</p>
        <p>Below, you will find samples of the images generated by StyleGAN3 for each minority class.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Results of Augmentation</title>
      </sec>
      <sec id="sec-4-3">
        <title>4.2.1. Approach 1: Traditional Augmentation</title>
        <p>Table 3 presents the performance of the ResNet50 and InceptionV3 models on the COVID-19
Radiography and HAM10000 datasets using Approach 1, evaluated in terms of accuracy, recall,
precision, and F1-score.</p>
        <p>It is observed that the InceptionV3 model slightly outperforms the ResNet50 model in terms of
accuracy and F1-score, achieving an accuracy of 94.82% compared to 91.26% for ResNet50 on the
COVID-19 dataset and 82.49% compared to 81.48% on the HAM10000 dataset
c) HAM10000 - ResNet50
Figure 2: Confusion matrices for ResNet50 and InceptionV3 on the COVID-19 and HAM10000
datasets with Approach 1.
d) HAM10000 - InceptionV3</p>
        <p>The confusion matrices in (see Figure 2) visualize the performance of the ResNet50 and
InceptionV3 models on the COVID-19 Radiography and HAM10000 datasets using the traditional
augmentation approach. They allow for evaluating the quality of each model’s predictions on these
two datasets in terms of correct and incorrect classifications.</p>
      </sec>
      <sec id="sec-4-4">
        <title>4.2.2. Approach 2: Traditional Data Augmentation with Balancing</title>
        <p>Table 4 presents the performance of the ResNet50 and InceptionV3 models on the COVID-19
Radiography and HAM10000 datasets using Approach 2, evaluated in terms of accuracy, recall,
precision, and F1-score.</p>
        <p>It is noteworthy that the InceptionV3 model slightly surpasses the ResNet50 model in terms of
accuracy and F1-score on the COVID-19 Radiography dataset, achieving an accuracy of 95.46%
compared to 94.33% for ResNet50. Conversely, for the HAM10000 dataset, it is the ResNet50 model
that displays better performance, achieving an accuracy of 83.19% compared to 71.77% for InceptionV3
c) HAM10000 - ResNet50
d) HAM10000 - InceptionV3</p>
      </sec>
      <sec id="sec-4-5">
        <title>4.2.3. Approach 3: Augmentation using StyleGAN3 with Batch Injection</title>
        <p>We observe that the InceptionV3 model slightly outperforms the ResNet50 model in terms of
accuracy and F1-score on the COVID-19 Radiography dataset, achieving an accuracy of 95.46%
compared to 94.33% for ResNet50. However, on the HAM10000 dataset, the InceptionV3 model also
achieves better results with an accuracy of 86.08% compared to 83.19% for ResNet50 (see Figure 4 for
the confusion matrices).</p>
      </sec>
      <sec id="sec-4-6">
        <title>4.2.4. Approach 4: Augmentation using StyleGAN3 and Balancing with</title>
      </sec>
      <sec id="sec-4-7">
        <title>Traditional Methods</title>
        <p>In this approach, we tested different levels of data augmentation, namely 10%, 20%, 30%, and 40%.</p>
      </sec>
      <sec id="sec-4-8">
        <title>4.2.5. With 10% Augmentation</title>
        <p>Table 6 presents the performance of the ResNet50 and InceptionV3 models on the COVID-19
Radiography and HAM10000 datasets with 10% data augmentation using Approach 4, evaluated in
terms of accuracy, recall, precision, and F1-score.</p>
        <p>It is observed that the InceptionV3 model continues to outperform the ResNet50 model in terms of
accuracy and F1-score with this approach on the COVID-19 Radiography dataset, achieving an
accuracy of 95.84% compared to 94.49% for ResNet50. Furthermore, on the HAM10000 dataset,
InceptionV3 also demonstrates better performance with an accuracy of 86.43% versus 83.99% for
ResNet50.</p>
      </sec>
      <sec id="sec-4-9">
        <title>4.2.6. With 20% Augmentation</title>
        <p>c) HAM10000 - ResNet50 d) HAM10000 - InceptionV3
Figure 6: Confusion matrices for ResNet50 and InceptionV3 on the COVID-19 and HAM10000
datasets with approach 4 at 30%.</p>
        <p>It can be observed that in the COVID-19 Radiography dataset, InceptionV3 outperforms ResNet50
with an accuracy of 95.58% compared to 93.76%. The F1 scores, recall, and precision further confirm
this trend, indicating better overall performance for InceptionV3.</p>
        <p>Regarding the HAM10000 dataset, although the gap is smaller, InceptionV3 maintains an
advantage with an accuracy of 85.73% compared to 84.88% for ResNet50. This difference demonstrates
better handling of complex classes by InceptionV3.</p>
      </sec>
      <sec id="sec-4-10">
        <title>4.2.7. With 30% Augmentation</title>
        <p>It can be noted that the InceptionV3 model continues to surpass the ResNet50 model in terms of
accuracy and F1-score with this approach on the COVID-19 Radiography dataset, achieving an
accuracy of 95.84% compared to 94.49% for ResNet50. Furthermore, on the HAM10000 dataset,
InceptionV3 also demonstrates better performance with an accuracy of 86.43% compared to 83.99%
for ResNet50.
c) HAM10000 - ResNet50 d) HAM10000 - InceptionV3
Figure 7: Confusion matrices for ResNet50 and InceptionV3 on the COVID-19 and HAM10000
datasets with approach 4 at 20%.</p>
      </sec>
      <sec id="sec-4-11">
        <title>4.2.8. With 40% augmentation</title>
        <p>The table 10 presents the performance of the InceptionV3 and ResNet50 models on the COVID-19
Radiography and HAM10000 datasets with a 40% data augmentation using approach 4.</p>
        <p>The results show that InceptionV3 continues to exhibit better performance compared to ResNet50
in terms of accuracy and F1-score. For the COVID-19 dataset, InceptionV3 achieves an accuracy of
95.18%, while ResNet50 shows an accuracy of 93.24%. Regarding the HAM10000 dataset, InceptionV3
achieves an accuracy of 85.38%, compared to 82.94% for ResNet50, confirming the effectiveness of
InceptionV3 for this approach.
c) HAM10000 - ResNet50 d) HAM10000 - InceptionV3
Figure 8: Confusion matrices for ResNet50 and InceptionV3 on the COVID-19 and HAM10000
datasets with approach 4 at 30%.</p>
      </sec>
      <sec id="sec-4-12">
        <title>4.3. Discussion and comparison</title>
        <p>For the COVID-19 Radiography dataset, approach 4, which includes a 30% increase in generated
data, proved to be the most effective. It achieved an accuracy of 95.96% for the InceptionV3 model and
94.49% for ResNet50, outperforming all other tested approaches.</p>
        <p>In the HAM10000 dataset, approach 4 also delivered the best results, especially for InceptionV3,
which reached an accuracy of 86.43%, surpassing other methods.</p>
        <p>Approach 1 relies on traditional data augmentation, but its limitations quickly become apparent.
On the COVID-19 Radiography dataset, it allows InceptionV3 to achieve an accuracy of 94.82% and
ResNet50 91.26%, while on the HAM10000 dataset, the performances are 83.09% for InceptionV3 and
82.59% for ResNet50. However, this unbalanced approach does not provide significant improvements
and may even lead to decreased performance due to the persistent class imbalance. This is where
approach 2, which incorporates class balancing along with data augmentation, proves to be more
effective. Indeed, on the COVID-19 Radiography dataset, it achieves an accuracy of 95.89% for
InceptionV3 and 94.40% for ResNet50, and on HAM10000, the results are also improved, with 86.03%
for InceptionV3 and 84.49% for ResNet50. This demonstrates that the addition of class balancing
allows for better model generalization, particularly on unbalanced datasets, making approach 2 more
effective than approach 1.</p>
        <p>In approach 3, although the generated images are of good quality, the main issue lies in the
constant variation of the images injected into the batches at each training step. This fluctuation
prevents the model from converging effectively, as it cannot adapt well to changing data. The lack of
consistency in the batches disrupts the model’s learning, limiting overall performance improvements.
In comparison, approach 2, which uses static data balancing, offers greater stability and enables the
model to converge better, thus explaining its superior results.
c) HAM10000 - ResNet50 d) HAM10000 - InceptionV3
Figure 9: Confusion matrices for ResNet50 and InceptionV3 on the COVID-19 and HAM10000
datasets with approach 4.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusion</title>
      <p>Conclusion Data augmentation is key to optimizing the performance of deep learning models in
medical imaging, especially in the presence of imbalanced and challenging datasets. In this work, the
application of StyleGAN in generating synthetic images was investigated and its impact on the
training of classification models evaluated.</p>
      <p>Our experiments demonstrated that while the synthetic images generated by StyleGAN are
realistic and useful for augmenting datasets, they alone are not sufficient to surpass the performance
obtained using traditional data augmentation methods. The diversity and realism of the generated
images remain a concern, especially with the complexity and variability of medical images, which are
not always well-modeled by generative models.</p>
      <p>However, this study shows the potential of GANs for enhancing medical classification datasets.
While ResNet50 and InceptionV3 have worked effectively, other architectures and fine-tuning
strategies can potentially improve model robustness.</p>
      <p>Prospective Pathways To supplement this study, there are numerous paths that can be taken:
 Hybrid methods: Combining GANs with other generation techniques, i.e., variational
autoencoders, would further improve synthetic data quality. Systematic clinical validation: Testing
these techniques in actual clinical environments to determine their feasibility.
 Combining imbalance handling techniques: Addressing dataset imbalance by integrating two
approaches, such as undersampling the majority classes through pruning while simultaneously
oversampling the minority classes, can improve model generalization and mitigate bias.</p>
    </sec>
    <sec id="sec-6">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the authors used Grammarly in order to: Grammar and spelling
check. After using this tool, the authors reviewed and edited the content as needed and take full
responsibility for the publication’s content.
Two-step training using actual and synthesized images by progressive growing of generative
adversarial networks, PLOS ONE 15 (2020) e0229951.
[12] A. Waheed, M. G. G. Khan, M. Ali, M. A. G. Javed, J. Liao, A. I. T. S., G. Yang, CovidGAN: Data
augmentation using auxiliary classifier GAN for improved COVID-19 detection, IEEE Access 8
(2020) 91916–91923.
[13] V. Sandfort, K. Yan, P. J. Pickhardt, R. M. Summers, Data augmentation using generative
adversarial networks (CycleGAN) to improve generalizability in CT segmentation tasks,
Scientific Reports 9 (2019) 16884.
[14] T. Karras, S. Laine, T. Aila, A style-based generator architecture for generative adversarial
networks, in: Proceedings of the IEEE/CVF conference on computer vision and pattern
recognition, 2019, pp. 4401–4410.
[15] T. Karras, S. Laine, M. Aittala, J. Hellsten, J. Lehtinen, T. Aila, Analyzing and improving the
image quality of stylegan, in: Proceedings of the IEEE/CVF conference on computer vision and
pattern recognition, 2020, pp. 8110–8119.
[16] T. Karras, M. Aittala, S. Laine, E. Härkönen, J. Hellsten, J. Lehtinen, T. Aila, Alias-free
generative adversarial networks, Advances in neural information processing systems 34 (2021)
852–863.
[17] A. Waheed, M. Goyal, D. Gupta, A. Khanna, F. Al-Turjman, P. R. Pinheiro, Covidgan: Data
augmentation using auxiliary classifier gan for improved covid-19 detection, IEEE Access 8
(2020) 91916–91923.
[18] S. V. J, J. F. D, Deep learning algorithm for covid-19 classification using chest x-ray images,</p>
      <p>Computational and Mathematical Methods in Medicine 2021 (2021) Article ID 9269173, 10 pages.
[19] B. Ahmad, S. Jun, V. Palade, Q. You, L. Mao, M. Zhongjie, Improving skin cancer classification
using heavy-tailed student t-distribution in generative adversarial networks (ted-gan),
Diagnostics 11 (2021) 2147.
[20] Q. Su, H. N. A. Hamed, M. A. Isa, X. Hao, X. Dai, A gan-based data augmentation method for
imbalanced multi-class skin lesion classification, IEEE Access 12 (2024) 16498–16513.
[21] A. Bilal, J. Sun, Q. You, V. Palade, Z. Mao, Brain tumor classification using a combination of
variational autoencoders and generative adversarial networks, Biomedicines 10 (2022) 1–19.
[22] G. C. Oliveira, G. H. Rosa, D. C. G. Pedronette, J. P. Papa, H. Kumar, L. A. Passos, D. Kumar,
Which generative adversarial network yields high-quality synthetic medical images:
Investigation using amd image datasets, arXiv preprint arXiv:2203.13856v1 (2022).
[23] P. Tschandl, C. Rosendahl, H. Kittler, The ham10000 dataset, a large collection of multi-source
dermatoscopic images of common pigmented skin lesions, Scientific data 5 (2018) 1–9.
[24] U. Rehman, T. COVID, Radiography database| kaggle, 19. URL:
https://www.kaggle.com/datasets/tawsifurrahman/covid19-radiography-database.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>F.</given-names>
            <surname>Garcea</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Serra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Lamberti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Morra</surname>
          </string-name>
          ,
          <article-title>Data augmentation for medical imaging: A systematic literature review</article-title>
          ,
          <source>Computers in Biology and Medicine</source>
          <volume>152</volume>
          (
          <year>2023</year>
          )
          <fpage>106391</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>W.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Yu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Shi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. P.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <article-title>A survey on imbalanced learning: latest research, applications and future directions</article-title>
          ,
          <source>Artificial Intelligence Review</source>
          <volume>57</volume>
          (
          <year>2024</year>
          )
          <fpage>137</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>P.</given-names>
            <surname>Kumar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Bhatnagar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Gaur</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Bhatnagar</surname>
          </string-name>
          ,
          <article-title>Classification of imbalanced data: review of methods and applications</article-title>
          , in: IOP conference series:
          <article-title>materials science and engineering</article-title>
          , volume
          <volume>1099</volume>
          ,
          <string-name>
            <given-names>IOP</given-names>
            <surname>Publishing</surname>
          </string-name>
          ,
          <year>2021</year>
          , p.
          <fpage>012077</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>S.</given-names>
            <surname>Yadav</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G. P.</given-names>
            <surname>Bhole</surname>
          </string-name>
          ,
          <article-title>Handling imbalanced dataset classification in machine learning</article-title>
          ,
          <source>in: 2020 IEEE Pune Section International Conference (PuneCon)</source>
          , IEEE,
          <year>2020</year>
          , pp.
          <fpage>38</fpage>
          -
          <lpage>43</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>M.</given-names>
            <surname>Buda</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Maki</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. A.</given-names>
            <surname>Mazurowski</surname>
          </string-name>
          ,
          <article-title>A systematic study of the class imbalance problem in convolutional neural networks</article-title>
          ,
          <source>Neural networks 106</source>
          (
          <year>2018</year>
          )
          <fpage>249</fpage>
          -
          <lpage>259</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>T.-Y.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Goyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Girshick</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>He</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Dollár</surname>
          </string-name>
          ,
          <article-title>Focal loss for dense object detection</article-title>
          ,
          <source>in: Proceedings of the IEEE international conference on computer vision</source>
          ,
          <year>2017</year>
          , pp.
          <fpage>2980</fpage>
          -
          <lpage>2988</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>F.</given-names>
            <surname>Touazi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Gaceb</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Chirane</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Herzalleh</surname>
          </string-name>
          ,
          <article-title>Two-stage approach for semantic image segmentation of breast cancer: Deep learning and mass detection in mammographic images</article-title>
          , in: International Conference on Informatics &amp;
          <string-name>
            <surname>Data-Driven Medicine</surname>
          </string-name>
          ,
          <year>2023</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>13</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>M.</given-names>
            <surname>Khaled</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Touazi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Gaceb</surname>
          </string-name>
          ,
          <article-title>Improving breast cancer diagnosis in mammograms with progres- sive transfer learning and ensemble deep learning</article-title>
          ,
          <source>Arabian Journal for Science and Engineering</source>
          (
          <year>2024</year>
          )
          <fpage>1</fpage>
          -
          <lpage>24</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>I. J.</given-names>
            <surname>Goodfellow</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Pouget-Abadie</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Mirza</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Warde-Farley</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Ozair</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Courville</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Bengio</surname>
          </string-name>
          ,
          <article-title>Generative adversarial nets</article-title>
          ,
          <source>in: Advances in Neural Information Processing Systems</source>
          <volume>27</volume>
          ,
          <year>2014</year>
          , pp.
          <fpage>2672</fpage>
          -
          <lpage>2680</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>J.</given-names>
            <surname>Islam</surname>
          </string-name>
          ,
          <string-name>
            <surname>Y. Zhang,</surname>
          </string-name>
          <article-title>GAN-based synthetic brain PET image generation</article-title>
          ,
          <source>Brain Informatics</source>
          <volume>7</volume>
          (
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>A.</given-names>
            <surname>Teramoto</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Tsukamoto</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Kiriyama</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Fujita</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Yamamoto</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Tsukiji</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Imaizumi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Toyama</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Oda</surname>
          </string-name>
          , T. Kudo,
          <article-title>Deep learning approach to classification of lung cytological images:</article-title>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>