<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>Y. Cheng, D. Wang, P. Zhou, T. Zhang, Model Compression and Acceleration for Deep Neural
Networks: The Principles, Progress, and Challenges, IEEE Signal Process. Mag.</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <article-id pub-id-type="doi">10.1109/msp.2017.2765695</article-id>
      <title-group>
        <article-title>Liminal efficiency for AI general models: investigating compression techniques for maximized performance in the glaucoma's diagnosis and prognosis⋆</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Volodymyr Kysil</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Olga Drachuk</string-name>
          <email>drachuk@vnmu.edu.ua</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Anatolii Korol</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Valentyna Hnenna</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Elena</string-name>
          <email>elena.zaitseva@fri.uniza.sk</email>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Khmelnytskyi National University</institution>
          ,
          <addr-line>Institutska str., 11, Khmelnytskyi, 29016</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>National Pirogov Memorial Medical University</institution>
          ,
          <addr-line>Pirogova str., 56, Vinnytsya, 21018</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Zilina University</institution>
          ,
          <addr-line>Univerzitná 8215, 010 26 Žilina</addr-line>
          ,
          <country country="SK">Slovakia</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2019</year>
      </pub-date>
      <volume>1</volume>
      <issue>2018</issue>
      <fpage>1</fpage>
      <lpage>11</lpage>
      <abstract>
        <p>Deep learning models have demonstrated remarkable performance across the glaucoma's diagnosis and prognosis; however, deploying them in resource-constrained environments poses significant challenges. This research explores the balance between compression and accuracy preservation in specialist convolutional neural networks (CNNs) intended for CPU-based execution with minimal storage requirements. By employing pruning, knowledge distillation, quantization, and weight sharing, it is aimed to achieve maximal compression without compromising essential task performance. Resulting findings provide insights into the efficiency limits of model compression and its implications for real-world deployment. Additionally, the applicability of these compression techniques to Transformer-based architectures is examined throughout the work, which pose unique challenges due to their reliance on attention mechanisms. deep learning, glaucoma's diagnosis and prognosis, medical imaging optic nerve segmentation, fundus image analysis, model compression, pruning, knowledge distillation, quantization, weight sharing, transformer models, convolutional neural networks (CNN), low-resource deployment, edge computing, self-attention mechanisms, transfer learning, efficiency, optimization, neural network performance, sparse models, inference speed, binary classification. 1 Intelitsis'25: The 6th International Workshop on Intelligent Information Technologies &amp; Systems of Information Security, April 04, 2025, Khmelnytskyi, Ukraine ∗ Corresponding author. † These authors contributed equally.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Glaucoma is the second most common cause of blindness and disability in the world [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. It affects
more than 90 million people worldwide [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. It is estimated that in 2013, the number of people aged
40-80 years suffering from glaucoma was 64.3 million, in 2020 - about 76.0 million, and by 2040 this
figure is projected to increase to 111.8 million [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ].
      </p>
      <p>The main difficulty in diagnosing glaucoma is that its symptoms usually appear only when vision
has already deteriorated significantly. Due to the asymptomatic course and slow development of the
disease, many people are unaware of the problem, which makes it difficult to detect and predict it
early, especially in the initial stages. However, timely diagnosis can slow down the progression of
the disease and prevent vision loss.</p>
      <p>
        Since glaucoma causes retinal damage due to damage to the optic nerve head and increased
intraocular pressure, an important step in its detection is the segmentation of the optic disc. This
procedure is difficult due to the small size of the disc and possible blood supply disorders [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ].
      </p>
      <p>
        The standard diagnosis of glaucoma includes an assessment of the optic nerve head by examining
the retina, measuring intraocular pressure, analyzing visual fields, and taking into account other
related factors [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. At the same time, traditional methods are time-consuming, largely dependent on
the ophthalmologist's experience, and may be accompanied by human error, which makes it difficult
to detect the disease in the early stages.
      </p>
      <p>
        A promising area for improving the diagnostic process is modern information consulting and
diagnostic technologies. They help to improve medical decisions and the quality of treatment of
patients with retinal pathologies and are already used in numerous systems for diagnosing various
diseases [
        <xref ref-type="bibr" rid="ref5 ref6 ref7 ref8">5-8</xref>
        ].
      </p>
      <p>
        Over the past decade, machine learning and deep learning methods have developed significantly,
not only improving the accuracy of glaucoma detection but also speeding up the processing of large
amounts of images and data, making the work of doctors easier [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. Studies confirm that the use of
these technologies allows for the automatic diagnosis of glaucoma in the early stages by analyzing
fundus and optical coherence tomography images to classify the eye as normal or glaucoma-affected
[
        <xref ref-type="bibr" rid="ref10 ref11">10, 11</xref>
        ].
      </p>
      <p>Recent advancements in deep learning have led to increasingly large and complex models, making
deployment on limited-resource environments such as edge devices and low-power CPUs
impractical. While specialist models tailored for specific tasks offer high accuracy, they often retain
redundant parameters that inflate model size and computational overhead. The challenge lies in
optimizing these models for real-world applications while maintaining their effectiveness.</p>
      <p>
        Compression techniques [
        <xref ref-type="bibr" rid="ref12 ref13 ref14">12-14</xref>
        ] such as pruning (removes unnecessary weights, reducing
computation and memory footprint), quantization (converts high-precision weights into lower-bit
representations, minimizing storage requirements and enhancing inference speed; applied to an
already-learned model for increasing performance speed), weight sharing (groups similar weights to
reduce parameter redundancy while preserving network expressiveness; also applied to an
alreadylearned model for increasing performance speed and decrease storage space), and knowledge
distillation (transfers knowledge from a larger, more complex model to a smaller, more efficient
model) can significantly reduce model size and inference time. However, an aggressive compression
strategy may result in degraded accuracy, raising the question: How far can a model be compressed
while maintaining usability?
      </p>
    </sec>
    <sec id="sec-2">
      <title>2. Related works</title>
      <p>
        The balance between model size and performance has been widely studied. Han et al. [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ] introduced
structured and unstructured pruning to eliminate redundant weights, demonstrating notable
efficiency improvements. Frankle &amp; Carbin [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ] proposed the Lottery Ticket Hypothesis, suggesting
that sparse subnetworks can match full model performance. Other works [17-19] explore
quantization and architectural optimizations to further enhance efficiency.
      </p>
      <p>Additionally, research by Bala and Eschbach [20] highlights the importance of optimized color
spaces for improving neural network visibility in medical imaging. Their findings inform our
preprocessing techniques for fundus image analysis.</p>
      <p>Studies such as Wu et al. [18] emphasize the delicate balance between compression level and
accuracy. Excessive pruning or quantization can introduce errors that degrade performance,
highlighting the need for a strategic approach to compression. Knowledge distillation has also been
explored as a means to retain essential knowledge in highly compressed models while mitigating
accuracy loss.</p>
      <p>Research has shown that specialized models tend to outperform general-purpose models in their
respective domains [21-24]. This motivated selection of an image transformer model for optic nerve
extraction, as transformers have demonstrated superior performance in medical imaging tasks. By
leveraging their self-attention mechanisms, these models excel in capturing fine-grained spatial
dependencies, which is crucial for precise segmentation of the optic nerve in fundus images.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Methodology</title>
      <sec id="sec-3-1">
        <title>3.1. Baseline model description</title>
        <p>CNN-based encoder-decoder architecture was utilized, structured as follows:
1. Initial feature extraction – VGG-16 backbone with four pooling layers.
2. Intermediate processing – custom layers for high-level feature abstraction.
3. Decoding – three decoder layers with skip connections to preserve spatial information.</p>
        <p>This model is fine-tuned for binary classification of optic nerve localization in fundus images.
Model is used for image segmentation into single-channel optic mask [25]. The model is tasked to
learn high-level features, which, in practice, may further decompose into finer feature
representations depending on the model's depth and granularity. Feature rough estimate is 17
discrete low-level features that the CNN must detect, process, and classify to correctly locate the
optic nerve. Feature breakdown follows:</p>
        <sec id="sec-3-1-1">
          <title>Feature 1: Bright round/oval spot (optic disc):</title>
          <p>1. Brightness intensity contrast (1).
2. Circular shape detection (1).
3. Local texture gradient (1).
4. Edge detection (1).
5. Color consistency (1).</p>
          <p>Feature 2: Conjunction of several blood vessels:
6. Vessel edge detection (1).
7. Vessel thickness estimation (1).
8. Vessel directionality (1).
9. Vessel junction detection (1).
10. Vessel curvature (1).</p>
          <p>Feature 3: Rounded dark border around the bright spot:
11. Contrast difference (1).
12. Border shape (1).
13. Edge sharpness (1).
14. Shadowing effects (1).</p>
          <p>Feature 4: Not on the black part of the image:
15. Background segmentation (1).
16. Foreground detection (1).
17. Image region classification (1).</p>
          <p>Image describing structure of base-64 transformer model is provided in Figure 1.</p>
        </sec>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Image preprocessing</title>
        <p>To enhance image quality and improve feature extraction, histogram equalization and CLAHE
(Contrast Limited Adaptive Histogram Equalization) in the LAB color space are applied. These
techniques help normalize contrast and highlight relevant structures in the fundus images. The
choice of the LAB color space is inspired by research from [26, 27], which demonstrates that
optimized color spaces can improve neural network visibility and feature extraction in medical
imaging.</p>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. Compression techniques</title>
        <p>Iterative structured pruning is being applied to model, progressively removing low-importance
weights while monitoring accuracy loss, resulting model was fine-tuned. In this article pruning was
pretty aggressive (50%). Dropout layers (30% rate) were also applied to force model to compress
knowledge and decrease amount of knowledge lost on pruning application, lessening retrain time
and shortening preparation for further compression.</p>
        <p>A smaller student model is trained to mimic the larger teacher model, learning its essential
features while significantly reducing parameter count. In this article larger first iteration model was
used as teacher model for all further iterations of student model learning.</p>
        <p>Post-training quantization was employed to convert 32-bit floating-point weights to 8-bit integers,
reducing memory consumption. Spoiler: memory consumption remained unchanged, inference time
remained largely unchanged (5% speed increase), space on disk actually increased (why?).</p>
        <p>K-Means clustering was used to group similar weights, replacing them with shared values to
reduce parameter storage. Spoiler: ineffective for accuracy, inference time unchanged, memory
consumption unchanged.</p>
        <p>Transfer learning was employed to teach the first large iterations of the transformer faster,
leveraging pre-trained models to accelerate convergence and improve performance. Prior research
has demonstrated the effectiveness of transfer learning in deep learning applications, particularly in
medical imaging tasks, where pre-trained architectures significantly enhance detection accuracy and
efficiency [28].</p>
      </sec>
      <sec id="sec-3-4">
        <title>3.4. Transformer model considerations</title>
        <p>In addition to the CNN-based approach detailed above, current research evaluates an image
transformer model for optic nerve extraction. Transformer architectures are renowned for their
selfattention mechanisms, which excel at capturing long-range dependencies and fine-grained
contextual relationships-a quality that is particularly beneficial for specialized tasks in medical
imaging. In the context of optic nerve segmentation, transformers are capable of discerning subtle
variations in image features that are crucial for accurate detection.</p>
        <p>While recent literature has introduced a range of advanced compression strategies tailored to
transformers (example - low-rank factorization [29, 30] and sparse attention mechanisms [31, 32]),
current work does not implement these methods. Instead, the transformer model is utilized in its
standard form to serve as a benchmark for specialized model performance. This approach allows to
directly compare the efficiency gains achieved through CNN compression techniques - pruning,
knowledge distillation, quantization, and weight sharing - with the inherent advantages of
transformer-based models operating under similar CPU constraints.</p>
        <p>By focusing on this comparison, we seek to determine where the transformer’s compressed
performance can match accuracy of base transformer model. This evaluation is critical to overall
theme of liminal efficiency, where we explore the threshold at which aggressive compression
techniques can be applied without incurring unacceptable losses in accuracy, while considering the
trade-offs between generalist and specialist model architectures. Liminal efficiency is defined as &lt;5%
accuracy loss from starting model without losing binary feature classification metrics and increasing
performance.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Experiments and testing</title>
      <p>For experiments we use datasets ODIR-5K [33] – local ophthalmologist-provided datasets, which
consists of 500 random images selected for training and evaluation.</p>
      <p>We calculated following metrics: inference time, total training time, number of training epochs
for sufficient results (here training epoch is one run over training image set), evaluation prediction
count, total true positive predictions, total true negative predictions, total false positive predictions,
total false negative predictions, additional false detections, model size, and complex metrics:
accuracy, recall, and F1-score.</p>
      <p>For testing the results 871 images tested in a CPU environment.</p>
      <p>Experiments are conducted on model bases of 64, 32, 16, 8. Iteration model graphic images for
reference of differences in iteration for base 32 and base 16 are displayed on Figure 2 and Figure 3
respectively (difference is in thickness of convolution blocks – these are multipliers of base).</p>
      <p>Training was conducted using k-fold cross-validation algorithm with resulting loss being
extracted from evaluation fold. Additionally, to preserve edge cases knowledge a set of highest loss
cases in maximum size of 1 fold was assembled and used every step to additionally randomly select
a batch to train upon using weighted random where higher loss cases are selected more often. Both
knowledge distillation and pruning were trained on similar edge cases.</p>
      <sec id="sec-4-1">
        <title>4.1. Pruning experiments</title>
        <p>Pruning experiment results are provided in Table 1. Pruning experiments involve following:
1. Record metrics for each iteration, epoch and training session.
2. At iteration 1 transfer-learning is applied to model encoder blocks 1 through 4, then dropout
is applied at each activation except the very last block using dropout-30% layers. After that
model is trained to find optic nerve until sufficient results are obtained.
3. At subsequent iterations: previous iteration model is pruned by 50% with dropout turned off,
dropout is turned back on after 1 training epoch and model is subsequently fine-tuned until
sufficient accuracy is obtained.
4. Repeat for following “bases” of transformer: 64 (initial), 32, 16, 8. We stopped at base 8
because model stopped converging fast enough requiring obnoxious amount of training time
(recorded results are results of first converged training, however last iteration did not
converge at all).</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Knowledge distillation</title>
        <p>Knowledge distillation experiment results are provided in Table 2. Knowledge distillation
experiments involve following:
1. Record metrics for each iteration, epoch and training session.
2. At iteration 1 transfer-learning is applied to model encoder blocks 1 through 4. After that
model is trained to find optic nerve until sufficient results are obtained.
3. At subsequent iterations: use results from iteration 1 as training target.
4. Repeat for following iteration “bases” of transformer: 64(initial, step 2), 32, 16, 8. I stopped at
base 8 because model(again) stopped converging fast enough, requiring obnoxious amount
of training time. This time base 8 model was trained for another 40 epochs till sufficient
accuracy.</p>
        <p>Resulting images that denote edge cases handling are displayed in Figures 4-8.</p>
      </sec>
      <sec id="sec-4-3">
        <title>4.3. Quantization testing</title>
        <sec id="sec-4-3-1">
          <title>Quantization testing experiments involve following: 1. Prepare final model from training for certain model base for quantization. 2. Train prepared model to adjust it for quantization and quantize when results are sufficient. 3. Evaluate performance.</title>
        </sec>
        <sec id="sec-4-3-2">
          <title>Results: gains after quantization are negligible. No useful results.</title>
        </sec>
      </sec>
      <sec id="sec-4-4">
        <title>4.4. Weight sharing</title>
        <p>Weight sharing testing involve following: Use final model from training for certain model base and
apply weight sharing, fine-tune and reapply weight sharing till accuracy is acceptable.</p>
        <p>Results: unusable, breaks detection. Training and weight sharing after each epoch did not improve
result. Applying weight sharing after every training did not improve resulting accuracy. Applying
weight sharing after several epochs passed did not improve accuracy. Issue with weight sharing was
that there were too many interconnected target clusters for good enough selection of optic nerve.
Issue example provided in Figure 9.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusions</title>
      <p>Analysis of experiments’ results shows that the x16 base dropout pruning variant is a liminal point
in terms of accuracy/performance ratio. It exhibits negligible accuracy loss compared to previous
results while being smaller in size and 5.8 times faster than the x64 base transformer. Moreover, it
retains edge case detection capability, indicating superior knowledge retention and fine-tuning
potential. This variant remains undertrained and can be fine-tuned further along with
unconstraining with setting dropout to 0. Alternatively, data from knowledge distillation experiments
suggest that the x8 base transformer has performed well in general cases. Therefore, an x8
nodropout pruned variant may be obtained from the x16 dropout 30 variant by pruning it to x8 base
and fine-tuning it with dropout set to 0 for maximum performance and knowledge retention.</p>
    </sec>
    <sec id="sec-6">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the authors used X-GPT-4o in order to: Grammar and spelling
check. After using this service, the authors reviewed and edited the content as needed and take full
responsibility for the publication’s content.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>K.</given-names>
            <surname>Allison</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Patel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Alabi</surname>
          </string-name>
          ,
          <article-title>Epidemiology of Glaucoma: The Past, Present, and Predictions for the Future</article-title>
          ,
          <source>Cureus</source>
          (
          <year>2020</year>
          ). doi:
          <volume>10</volume>
          .7759/cureus.11686.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>Y.-C.</given-names>
            <surname>Tham</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T. Y.</given-names>
            <surname>Wong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H. A.</given-names>
            <surname>Quigley</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Aung</surname>
          </string-name>
          , C.-Y. Cheng,
          <source>Global Prevalence of Glaucoma and Projections of Glaucoma Burden through 2040, Ophthalmology</source>
          <volume>121</volume>
          .11 (
          <year>2014</year>
          )
          <fpage>2081</fpage>
          -
          <lpage>2090</lpage>
          . doi:
          <volume>10</volume>
          .1016/j.ophtha.
          <year>2014</year>
          .
          <volume>05</volume>
          .013.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>J. B.</given-names>
            <surname>Jonas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Aung</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. R.</given-names>
            <surname>Bourne</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. M.</given-names>
            <surname>Bron</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Ritch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Panda-Jonas</surname>
          </string-name>
          , Glaucoma, Lancet
          <volume>390</volume>
          .10108 (
          <year>2017</year>
          )
          <fpage>2183</fpage>
          -
          <lpage>2193</lpage>
          . doi:
          <volume>10</volume>
          .1016/s0140-
          <volume>6736</volume>
          (
          <issue>17</issue>
          )
          <fpage>31469</fpage>
          -
          <lpage>1</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>X.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. R.</given-names>
            <surname>Islam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Akter</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Ahmed</surname>
          </string-name>
          , E. Kazami,
          <string-name>
            <given-names>H. A.</given-names>
            <surname>Serhan</surname>
          </string-name>
          ,
          <string-name>
            <surname>A.</surname>
          </string-name>
          <article-title>Abd-alrazaq, S. Yousefi, Artificial intelligence in glaucoma: opportunities, challenges, and future directions</article-title>
          ,
          <source>Biomed. Eng. OnLine 22</source>
          .1 (
          <year>2023</year>
          ).
          <source>doi:10.1186/s12938-023-01187-8.</source>
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>T.</given-names>
            <surname>Hovorushchenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Herts</surname>
          </string-name>
          , Ye. Hnatchuk,
          <article-title>Concept of Intelligent Decision Support System in the Legal Regulation of the Surrogate Motherhood</article-title>
          .
          <source>CEUR-WS</source>
          <volume>2488</volume>
          (
          <year>2019</year>
          )
          <fpage>57</fpage>
          -
          <lpage>68</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>T.</given-names>
            <surname>Hovorushchenko</surname>
          </string-name>
          , Ye. Hnatchuk,
          <string-name>
            <given-names>A.</given-names>
            <surname>Herts</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Moskalenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Osyadlyi</surname>
          </string-name>
          ,
          <article-title>Theoretical and Applied Principles of Information Technology for Supporting Medical Decision-Making Taking into Account the Legal Basis</article-title>
          ,
          <source>CEUR-WS</source>
          <volume>3038</volume>
          (
          <year>2021</year>
          )
          <fpage>172</fpage>
          -
          <lpage>181</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Hnatchuk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Hovorushchenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Pavlova</surname>
          </string-name>
          ,
          <article-title>Methodology for the development and application of clinical decisions support information technologies with consideration of civillegal grounds</article-title>
          ,
          <source>Radioelectron. Comput. Syst. № 1</source>
          (
          <year>2023</year>
          )
          <fpage>33</fpage>
          -
          <lpage>44</lpage>
          . doi:
          <volume>10</volume>
          .32620/reks.
          <year>2023</year>
          .
          <volume>1</volume>
          .03.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>T.</given-names>
            <surname>Hovorushchenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Herts</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Hnatchuk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Sachenko</surname>
          </string-name>
          ,
          <article-title>Supporting the Decision-Making About the Possibility of Donation and Transplantation Based on Civil Law Grounds</article-title>
          ,
          <source>in: Advances in Intelligent Systems and Computing</source>
          , Springer International Publishing, Cham,
          <year>2020</year>
          , p.
          <fpage>357</fpage>
          -
          <lpage>376</lpage>
          . doi:
          <volume>10</volume>
          .1007/978-3-
          <fpage>030</fpage>
          -54215-3_
          <fpage>23</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>J. C.</given-names>
            <surname>Mathew</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Ilango</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Asha</surname>
          </string-name>
          ,
          <source>Machine Learning Techniques, Detection and Prediction of Glaucoma- A Systematic Review, Int. J. Recent Innov. Trends Comput. Commun</source>
          .
          <volume>11</volume>
          .
          <string-name>
            <surname>5s</surname>
          </string-name>
          (
          <year>2023</year>
          )
          <fpage>283</fpage>
          -
          <lpage>309</lpage>
          . doi:
          <volume>10</volume>
          .17762/ijritcc.v11i5s.
          <fpage>6655</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>V.</given-names>
            <surname>Kysil</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Popov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Drachuk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Hnenna</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Martyniuk</surname>
          </string-name>
          ,
          <source>Concept of Information Technology for Diagnosis and Prognosis of Glaucoma Based on Machine Learning Methods. CEUR-WS</source>
          .
          <volume>3675</volume>
          (
          <year>2024</year>
          )
          <fpage>171</fpage>
          -
          <lpage>181</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>T.</given-names>
            <surname>Hovorushchenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Moskalenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Osyadlyi</surname>
          </string-name>
          ,
          <article-title>Methods of medical data management based on blockchain technologies</article-title>
          ,
          <source>J. Reliab. Intell. Environ. 9.1</source>
          (
          <issue>2022</issue>
          )
          <fpage>5</fpage>
          -
          <lpage>16</lpage>
          . doi:
          <volume>10</volume>
          .1007/s40860-022- 00178-1.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>A.</given-names>
            <surname>Saxena</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. K.</given-names>
            <surname>Bishwas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. A.</given-names>
            <surname>Mishra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Armstrong</surname>
          </string-name>
          ,
          <article-title>Comprehensive Study on Performance Evaluation and Optimization of Model Compression: Bridging Traditional Deep Learning</article-title>
          and
          <source>Large Language Models</source>
          ,
          <year>2024</year>
          . URL: https://arxiv.org/pdf/2407.15904.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>E.</given-names>
            <surname>Dupuis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Novo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I. O</given-names>
            <surname>'Connor</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Bosio</surname>
          </string-name>
          ,
          <article-title>Sensitivity Analysis and Compression Opportunities in DNNs Using Weight Sharing</article-title>
          ,
          <source>in: 2020 23rd International Symposium on Design and Diagnostics of Electronic Circuits &amp; Systems (DDECS)</source>
          , IEEE,
          <year>2020</year>
          . doi:
          <volume>10</volume>
          .1109/ddecs50862.
          <year>2020</year>
          .
          <volume>9095658</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Na</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H. H.</given-names>
            <surname>Kim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. B.</given-names>
            <surname>Yoo</surname>
          </string-name>
          ,
          <article-title>Shared knowledge distillation for robust multi‐scale super‐resolution networks, Electron</article-title>
          . Lett. (
          <year>2022</year>
          ). doi:
          <volume>10</volume>
          .1049/ell2.
          <fpage>12526</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>S.</given-names>
            <surname>Han</surname>
          </string-name>
          ,
          <string-name>
            <surname>J</surname>
          </string-name>
          . Pool,
          <string-name>
            <given-names>J.</given-names>
            <surname>Tran</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W. J.</given-names>
            <surname>Dally</surname>
          </string-name>
          ,
          <article-title>Learning both weights and connections for efficient neural networks</article-title>
          ,
          <year>2015</year>
          . URL: https://arxiv.org/abs/1506.02626.
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>J.</given-names>
            <surname>Frankle</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Carbin</surname>
          </string-name>
          ,
          <article-title>The lottery ticket hypothesis: Finding sparse</article-title>
          ,
          <source>trainable neural networks</source>
          ,
          <year>2019</year>
          . URL: https://arxiv.org/abs/
          <year>1803</year>
          .03635.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>