<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>High Cost, Low Trust? MSA-PNet Fixes Both for Medical Imaging</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Dost Muhammad</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Muhammad Salman</string-name>
          <email>salmanuom04@gmail.com</email>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Malika Bendechache</string-name>
          <email>malika.bendechache@universityofgalway.ie</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>ADAPT Research Centre, School of Computer Science, University of Galway</institution>
          ,
          <country country="IE">Ireland</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>CRT-AI and ADAPT Research Centres, School of Computer Science, University of Galway</institution>
          ,
          <country country="IE">Ireland</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Department of Software Engineering, University of Malakand</institution>
          ,
          <country country="PK">Pakistan</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Deep learning (DL) has advanced medical ultrasound imaging by enabling automatic detection of subtle pathological features, notably in breast cancer diagnostics. However, mainstream architectures namely EficientNetB7 and VGG19 remain limited by high computational complexity and poor model explainability, which hinders their integration into clinical workflows. This paper proposes MSA-PNet-a Multi-Scale Attention-Enhanced Prototype Network-designed to provide eficient, accurate, and explainable ultrasound-based disease prediction. MSA-PNet introduces an adaptive Feature Pyramid Network (FPN) with learnable scale-aware fusion to capture discriminative features across variable spatial resolutions. A spatial attention module selectively enhances diagnostically relevant regions, while an auxiliary ROI segmentation branch produces spatially aligned tumour masks, reinforcing both localisation accuracy and clinical coherence. For transparency, MSA-PNet incorporates a prototype-based explainability module optimized via metric learning, enabling predictions to be grounded in classspecific prototypical patterns and visual reasoning. Comprehensive evaluations on the BUSI ultrasound dataset demonstrate the superiority of MSA-PNet over state-of-the-art baselines. It achieves a mean Dice coeficient of 79.92%, Jaccard index of 81.07%, and a Hausdorf distance of 26.14, significantly outperforming both EficientNetB7 and VGG19 across all metrics. Furthermore, MSA-PNet reduces inference time to 21.63 seconds-representing a 5× improvement in computational eficiency-making it highly suitable for real-time diagnostic deployment. By integrating multi-scale attention, prototype-based reasoning, and ROI-aware localisation into a unified architecture, MSA-PNet delivers not only robust diagnostic performance but also high-quality, clinically meaningful explanations. Its outputs exhibit strong alignment with expert-annotated tumour boundaries, thus enhancing trust, explainability, and applicability in high-stakes medical imaging environments. This framework represents a promising step toward the practical deployment of XAI systems in ultrasound-based diagnostics.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Eficient Explainable AI</kwd>
        <kwd>Attention-Guided Tumour Localisation</kwd>
        <kwd>Prototype-Driven Explanation</kwd>
        <kwd>Metric-Based Explainability</kwd>
        <kwd>XAI in Healthcare</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Medical ultrasound imaging plays a critical role in the early detection and diagnosis of life-threatening
diseases namely breast cancer, owing to its real-time capability, safety, and accessibility. However, the
interpretation of ultrasound images remains a challenging task due to their inherently low contrast,
high speckle noise, and variability in lesion morphology [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. Recent advances in deep learning (DL) have
significantly enhanced diagnostic performance in ultrasound imaging, enabling automated identification
of subtle pathological features that may elude human observation[
        <xref ref-type="bibr" rid="ref2">2</xref>
        ].
      </p>
      <p>
        While DL models namely EficientNetB7 [
        <xref ref-type="bibr" rid="ref3 ref4">3, 4</xref>
        ] and VGG19 [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] have demonstrated remarkable success
in medical imaging tasks, their deployment in real-time clinical workflows is constrained by two key
limitations: high computational cost and lack of explainability [
        <xref ref-type="bibr" rid="ref6 ref7">6, 7</xref>
        ]. The computational demands of
these models make them impractical for point-of-care settings and edge devices. More importantly, their
opaque decision-making processes undermine clinical trust, limiting their integration into diagnostic
practice [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ].
      </p>
      <p>
        Explainable Artificial Intelligence (XAI) has emerged as a vital field to address these challenges.
Techniques like Grad-CAM, LIME, and SHAP attempt to generate post-hoc visual explanations by
highlighting salient input regions [
        <xref ref-type="bibr" rid="ref10 ref9">9, 10</xref>
        ]. However, these methods often produce difuse or inconsistent
heatmaps and are rarely aligned with annotated pathological structures, particularly in ultrasound
images. Additionally, the field lacks standardized, quantitative metrics to rigorously assess explanation
quality and reliability, further hindering adoption in clinical domains [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ].
      </p>
      <p>To address these limitations, we propose the Multi-Scale Attention-Enhanced Prototype Network
(MSA-PNet)—a novel framework tailored for eficient, explainable, and clinically meaningful ultrasound
diagnosis. MSA-PNet incorporates an adaptive Feature Pyramid Network (FPN) that dynamically fuses
multi-resolution feature maps, guided by spatial attention modules to enhance the representation
of diagnostically relevant regions. This allows the model to robustly capture heterogeneous lesion
characteristics across multiple scales.</p>
      <p>Moreover, MSA-PNet integrates a prototype-based explainability mechanism trained via metric
learning, enabling the network to reason about new cases by comparing them to learned prototypical
examples. This facilitates case-based explainability aligned with clinical intuition. To further strengthen
explainability, we incorporate an explicit Region of Interest (ROI) localisation branch that highlights
tumour boundaries with high spatial precision, providing visual justifications closely aligned with
radiologist expectations.</p>
      <p>
        The contributions of this work are summarized as follows:
• We present MSA-PNet, a novel architecture that combines adaptive multi-scale attention and
prototype-driven reasoning to support eficient and explainable ultrasound-based diagnosis.
• Our model integrates an ROI segmentation head that provides fine-grained, interpretable tumour
localisation, promoting transparency and clinical usability.
• MSA-PNet outperforms state-of-the-art models in prediction accuracy, explainability, and
inference eficiency—achieving up to 5 × faster inference time compared to EficientNetB7 and
VGG19, while producing explanations that closely align with annotated clinical ground truth, as
quantitatively validated using Dice coeficient [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ], Jaccard index [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ], and Hausdorf distance
[
        <xref ref-type="bibr" rid="ref13">13</xref>
        ].
      </p>
      <p>The remainder of this paper is structured as follows: Section 2 outlines the materials and methodology
employed in this study. Section 3 presents the experimental results, followed by an in-depth discussion
in Section 4. Finally, Section 5 concludes the paper by summarizing the key findings and outlining
future research directions.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Materials and Methods</title>
      <sec id="sec-2-1">
        <title>2.1. Dataset</title>
        <p>
          This study utilises the publicly available Breast Ultrasound Images (BUSI) dataset [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ], which comprises
780 greyscale ultrasound images categorized into three diagnostic classes: benign, malignant, and normal.
Each image is paired with a corresponding binary mask annotated by clinical experts, indicating the
precise tumour region. The dataset presents a realistic clinical scenario for evaluating prediction,
segmentation, and explainability in breast ultrasound diagnosis.
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Preprocessing</title>
        <p>
          All images and masks were resized to a uniform resolution of 224 × 224 pixels to ensure consistency
across input dimensions. Pixel intensity values were normalized to the range [
          <xref ref-type="bibr" rid="ref1">0, 1</xref>
          ] to improve numerical
stability during training [
          <xref ref-type="bibr" rid="ref15 ref16">15, 16</xref>
          ]. To mitigate overfitting and enhance model generalisation, common
data augmentation techniques—including horizontal flips, rotations, and random crops—were applied
[
          <xref ref-type="bibr" rid="ref17 ref18">17, 18</xref>
          ]. The dataset was randomly split into training (80%) and validation (20%) sets for performance
evaluation.
        </p>
      </sec>
      <sec id="sec-2-3">
        <title>2.3. Implementation Environment</title>
        <p>The experiments were conducted using Python, chosen for its versatility and comprehensive ecosystem
of DL libraries. Model training and evaluation were performed on a computational system featuring an
AMD Ryzen 7 5700X eight-core processor and a 16GB NVIDIA GeForce RTX 4080 GPU, providing the
necessary computational power for eficient execution of DL workloads.</p>
      </sec>
      <sec id="sec-2-4">
        <title>2.4. Proposed Approach: MSA-PNet</title>
        <p>
          The proposed Multi-Scale Attention-Enhanced Prototype Network (MSA-PNet) is a unified DL
architecture designed to enhance both the diagnostic accuracy and explainability of ultrasound-based tumour
prediction. The network begins with a convolutional neural network (CNN) [
          <xref ref-type="bibr" rid="ref19 ref20">19, 20</xref>
          ] backbone to extract
hierarchical feature representations from input images. These features are extracted across multiple
layers, capturing increasingly abstract information, and are defined recursively as follows:
 = (−1 ),  ∈ {2, 3, 4, 5}
(1)
where  represents the output feature map at the -th stage, and (·) denotes the corresponding
convolutional operations.
        </p>
        <p>To enable multi-scale analysis, the outputs {2, 3, 4, 5} are first passed through 1 × 1 lateral
convolutions to project them into a common feature space:</p>
        <p>= Conv1×1 (),  ∈ {2, 3, 4, 5}
These projected features are then fused using an adaptive top-down pathway with learned scalar weights
 to modulate the contribution from each level:</p>
        <p>=  · Conv 3×3 ( + Upsample(+1))
This adaptive fusion enables the network to emphasize the most diagnostically informative scales for
lesion detection.</p>
        <p>To further refine the spatial context, a spatial attention mechanism is applied to the lowest-level
fused feature map 2. An attention map  is computed using a 7 × 7 convolution followed by a sigmoid
activation:
 = (Conv 7×7 (2))</p>
        <p>2* = 2 ⊙ 
This attention map is used to modulate 2 via element-wise multiplication, yielding an enhanced feature
map  2* :
This step ensures that subsequent computations are focused on spatial regions of diagnostic significance.</p>
        <p>To provide explicit localisation of tumour regions, an auxiliary ROI segmentation head processes  2*
through a stack of convolutional layers with ReLU and sigmoid activations. The predicted mask  is
computed as:</p>
        <p>=  (Conv 1×1 (ReLU(Conv3×3 ( 2* ))))
This mask highlights candidate lesion areas and contributes to both segmentation accuracy and
interpretability.</p>
        <p>Beyond segmentation, the model incorporates a prototype-based reasoning mechanism for transparent
classification. A set of  learnable prototypes { }=1 in R are used to represent class-specific feature

patterns. For a given input, the flattened feature representation  ∈ R×  is compared to each
prototype using cosine similarity:
sim(,  ) =</p>
        <p>⊤
‖ ‖ · ‖  ‖
(2)
(3)
(4)
(5)
(6)
(7)
 =</p>
        <p>∈top-

1 ∑︁ sim(,  )
 =  ·  +  
To evaluate the reliability of explainability generated heatmaps applied to the considered dataset, we
employed three key metrics: DC, JI, and HD. These metrics are will assess the alignment between
generated explanation heatmaps, and ground truth annotations provided by radiologist, ensuring that
the explanations are clinically relevant and explainable.
where  and  are learnable parameters. This formulation not only provides discriminative predictions
but also grounds them in semantically meaningful prototypes, facilitating clinical interpretation and
visual traceability.</p>
      </sec>
      <sec id="sec-2-5">
        <title>2.5. Evaluation Metrics for XAI</title>
        <p>To reduce sensitivity to local noise and spatial variation, similarity scores are averaged across the top-
highest-scoring spatial locations:
The resulting vector  = [1, . . . ,  ] encodes the degree of alignment between the input and each
prototype. Finally, class logits are computed using a linear classifier:
(8)
(9)
(10)
(11)
(12)</p>
        <sec id="sec-2-5-1">
          <title>2.5.1. Dice Coeficient (DC))</title>
          <p>
            The DC [
            <xref ref-type="bibr" rid="ref21 ref22">21, 22</xref>
            ] quantifies the overlap between two sets of pixels, typically the XAI-generated
explanation () and the ground truth (). It is defined as in Equation 10:
(, ) =
2| ∩ |
|| + ||
          </p>
          <p>Here, | ∩ | represents the number of overlapping pixels between the two sets, while || and ||
denote the total number of pixels in each set, respectively. The DC ranges from 0 (no overlap) to 1
(perfect overlap), making it an intuitive measure of similarity.</p>
        </sec>
        <sec id="sec-2-5-2">
          <title>2.5.2. Jaccard Index (JI)</title>
          <p>
            The JI [
            <xref ref-type="bibr" rid="ref12 ref23">23, 12</xref>
            ], also known as Intersection over Union (IoU), measures the ratio of the intersection to
the union of two sets. It is given in Equation 11:
 (, ) = | ∩ |
| ∪ |
Here, | ∪ | = || + || − | ∩ |
          </p>
          <p>represents the total number of unique pixels in either set. Like
DC, the JI ranges from 0 to 1 but is more sensitive to diferences in smaller regions, which is critical for
detecting subtle discrepancies in medical images.</p>
        </sec>
        <sec id="sec-2-5-3">
          <title>2.5.3. Hausdorf Distance (HD)</title>
          <p>
            The HD [
            <xref ref-type="bibr" rid="ref13">13</xref>
            ] evaluates the maximum distance between the boundary points of two sets, providing a
robust measure of how far the predicted explanation is from the ground truth. It is mathematically
expressed in Equation 12:
          </p>
          <p>︂{
(, ) = max sup inf (, ), sup inf (, )
∈ ∈
∈ ∈
︂}</p>
          <p>In this equation, (, ) represents the Euclidean distance between points  and , sup denotes the
maximum distance over all points in the set, and inf represents the minimum distance to the closest
point in the other set. The Hausdorf Distance is particularly useful for evaluating boundary accuracy,
capturing the worst-case discrepancy between the edges of predicted and ground truth regions.</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Results and Analysis</title>
      <p>
        The diagnostic performance and explainability of the proposed MSA-PNet framework were
assessed through a comprehensive comparative study involving two widely adopted DL architectures:
EficientNetB7 [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] and VGG19 [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. Experiments were carried out using the BUSI ultrasound dataset,
employing evaluation metrics namely the Dice coeficient, Jaccard index, Hausdorf distance, and overall
computational inference time.
      </p>
      <p>Quantitative results are summarized in Table 1. MSA-PNet achieved a significantly higher mean
Dice coeficient (79.92% ± 1.75%) and Jaccard index (81.07% ± 1.53%) compared to EficientNetB7 (48.27%
Dice, 49.11% Jaccard) and VGG19 (46.43% Dice, 47.07% Jaccard). These metrics indicate superior spatial
overlap between the predicted and ground truth tumor regions, demonstrating MSA-PNet’s enhanced
ability to delineate lesion boundaries with greater precision.</p>
      <p>Furthermore, the Hausdorf distance—measuring the maximum deviation between predicted and
annotated contours—was substantially lower for MSA-PNet (26.14 ± 9.71) compared to EficientNetB7
(237.52 ± 58.90) and VGG19 (256.82 ± 77.24), reflecting finer edge localisation and more anatomically
consistent segmentations. Importantly, MSA-PNet required only 21.63 seconds to process the complete
validation set, representing a five-fold improvement in inference eficiency over EficientNetB7 (107.46 s)
and VGG19 (108.43 s). These results confirm MSA-PNet’s computational scalability and clinical viability.</p>
      <p>In addition to the numerical results, Figure 1 presents a qualitative comparison of the explanation
heatmaps generated by each model. EficientNetB7 and VGG19 tend to produce broad, difuse saliency
regions that often extend beyond the pathological areas, lacking precise anatomical alignment. In
contrast, MSA-PNet consistently generates tight, well-localized activation maps that closely correspond
to the tumour regions, thereby enhancing the clinical interpretability of its predictions. The improved
spatial focus of MSA-PNet’s heatmaps stems from its built-in ROI segmentation branch and
prototypebased decision logic, both of which explicitly encode spatial alignment with known tumour patterns.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Discussion</title>
      <p>The empirical performance of MSA-PNet can be directly attributed to a series of integrated architectural
innovations, each addressing a known limitation of prior DL models in medical ultrasound imaging.
First, traditional CNNs namely VGG19 or EficientNetB7 operate with fixed-scale feature hierarchies,
making them suboptimal for capturing ultrasound-specific pathology, which can present with highly
variable lesion sizes, shapes, and textures. MSA-PNet overcomes this limitation through its adaptive
Feature Pyramid Network (FPN), which employs learnable fusion weights (Equation 3) to dynamically
prioritize spatial resolutions most relevant to the diagnostic task.</p>
      <p>Second, generic CNN models often rely solely on the global receptive field of their deeper layers
for context aggregation, which can result in the loss of fine-grained spatial cues essential for accurate
tumour localisation. MSA-PNet’s spatial attention mechanism (Equation 4) reweights the low-level
feature maps, forcing the model to selectively attend to diagnostically significant regions while
suppressing background noise. This attention-driven localisation improves the focus and precision of both
classification and segmentation outputs.</p>
      <p>Third, the inclusion of an explicit ROI segmentation head (Equation 6) allows MSA-PNet to generate
clinically actionable binary masks without requiring a separate post-processing pipeline. This auxiliary
output not only strengthens the spatial alignment of predicted tumour regions but also contributes to
overall model robustness during training via multi-task learning.</p>
      <p>Most notably, MSA-PNet introduces a novel prototype-based explainability module (Equations 7–9).
Unlike post-hoc methods such as Grad-CAM or LIME, which provide approximate and potentially
unstable explanations, the prototype mechanism grounds each classification decision in a concrete,
learned visual concept. These prototypes function as case-based reasoning anchors—comparing each
input with stored representations of prototypical benign, malignant, or normal patterns—ofering
clinicians a more transparent and traceable decision path.</p>
      <p>The reduction in Hausdorf distance (over 90% improvement compared to baselines) demonstrates that
MSA-PNet excels not only in semantic prediction but also in precise boundary localisation. Moreover, the
significant drop in inference time underscores the model’s suitability for real-time clinical deployment,
an essential requirement for diagnostic imaging systems in fast-paced environments.</p>
      <p>Together, these innovations enable MSA-PNet to bridge the long-standing gap between model
performance and clinical trust. It delivers explainable predictions with quantitative and qualitative
alignment to radiologist expectations—unlike traditional black-box models whose outputs are dificult to
validate or interpret. These results strongly position MSA-PNet as a practical, accurate, and trustworthy
solution for AI-assisted ultrasound diagnostics.</p>
      <p>Beyond technical performance, the proposed MSA-PNet framework holds substantial clinical impact
by addressing core barriers to AI adoption in medical imaging—namely trust, transparency, and
deployment eficiency. By delivering fast, accurate, and explainable predictions, this approach has the potential
to enhance radiologist workflow, support early cancer detection, and extend diagnostic capabilities to
resource-constrained settings where expert availability is limited. Thus, MSA-PNet not only advances
the state of the art in XAI but also lays a practical foundation for integrating explainable DL models
into real-world clinical practice.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusion</title>
      <p>This study introduced MSA-PNet, a novel DL framework tailored for explainable and eficient
ultrasoundbased disease diagnosis. By integrating adaptive multi-scale attention mechanisms, a dedicated ROI
localisation module, and a prototype-based explainability layer, MSA-PNet addresses key limitations of
conventional DL models in terms of computational eficiency, spatial precision, and clinical transparency.
The model consistently outperformed state-of-the-art architectures namely EficientNetB7 and VGG19,
achieving higher diagnostic accuracy, more precise tumour localisation, and substantially reduced
inference time.</p>
      <p>In addition to its predictive capabilities, MSA-PNet provides visually and quantitatively robust
explanation maps that closely align with expert-annotated ground truths. These explainable outputs
enhance model trustworthiness and support clinical decision-making by ofering clear justifications for
each prediction. The model’s computational eficiency further facilitates real-time application in clinical
workflows, including deployment in low-resource settings. Overall, MSA-PNet not only advances
the methodological landscape of XAI in medical imaging but also demonstrates strong potential for
real-world integration in diagnostic radiology, improving both speed and reliability in ultrasound-based
disease detection.</p>
    </sec>
    <sec id="sec-6">
      <title>Acknowledgments</title>
      <p>This research was supported by Taighde Éireann – Research Ireland under grant numbers
GOIPG/2025/8471, 18/CRT/6223 (RI Centre for Research Training in Artificial Intelligence),
13/RC/2106/ _2 (ADAPT Centre),13/RC/2094/ _2 (Lero Centre) and College of Science and
Engineering, University of Galway. For the purpose of Open Access, the author has applied a CC BY public
copyright licence to any Author Accepted Manuscript version arising from this submission.</p>
    </sec>
    <sec id="sec-7">
      <title>Declaration on Generative AI</title>
      <p>OpenAI’s ChatGPT, Grammarly etc, were not used in the preparation of this manuscript. All content,
analysis, and writing were entirely conceived, developed, and validated by the authors.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>C. A.</given-names>
            <surname>Duarte-Salazar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. E.</given-names>
            <surname>Castro-Ospina</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. A.</given-names>
            <surname>Becerra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Delgado-Trejos</surname>
          </string-name>
          ,
          <article-title>Speckle noise reduction in ultrasound images for improving the metrological evaluation of biomedical applications: an overview</article-title>
          ,
          <source>IEEE Access 8</source>
          (
          <year>2020</year>
          )
          <fpage>15983</fpage>
          -
          <lpage>15999</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>D.</given-names>
            <surname>Muhammad</surname>
          </string-name>
          , I. Ahmed,
          <string-name>
            <given-names>M. O.</given-names>
            <surname>Ahmad</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Bendechache</surname>
          </string-name>
          ,
          <article-title>Randomized explainable machine learning models for eficient medical diagnosis</article-title>
          ,
          <source>IEEE Journal of Biomedical and Health Informatics</source>
          (
          <year>2024</year>
          )
          <fpage>1</fpage>
          -
          <lpage>10</lpage>
          . doi:
          <volume>10</volume>
          .1109/JBHI.
          <year>2024</year>
          .
          <volume>3491593</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>M.</given-names>
            <surname>Tan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Le</surname>
          </string-name>
          , Eficientnet:
          <article-title>Rethinking model scaling for convolutional neural networks</article-title>
          , arXiv preprint arXiv:
          <year>1905</year>
          .
          <volume>11946</volume>
          (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>D.</given-names>
            <surname>Muhammad</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Salman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Keles</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Bendechache</surname>
          </string-name>
          ,
          <article-title>All diagnosis: Can eficiency and transparency coexist? an explainable deep learning approach</article-title>
          ,
          <source>Scientific Reports</source>
          <volume>15</volume>
          (
          <year>2025</year>
          )
          <fpage>12812</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>K.</given-names>
            <surname>Simonyan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Zisserman</surname>
          </string-name>
          ,
          <article-title>Very deep convolutional networks for large-scale image recognition</article-title>
          ,
          <source>arXiv preprint arXiv:1409.1556</source>
          (
          <year>2014</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>D.</given-names>
            <surname>Muhammad</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Bendechache</surname>
          </string-name>
          ,
          <article-title>Can ai be faster, accurate, and explainable? spikenet makes it happen</article-title>
          ,
          <source>in: Annual Conference on Medical Image Understanding and Analysis</source>
          , Springer,
          <year>2025</year>
          , pp.
          <fpage>43</fpage>
          -
          <lpage>57</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>M.</given-names>
            <surname>Ali</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Muhammad</surname>
          </string-name>
          ,
          <string-name>
            <surname>O. I. Khalaf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Habib</surname>
          </string-name>
          ,
          <article-title>Optimizing mobile cloud computing: A comparative analysis and innovative cost-eficient partitioning model</article-title>
          ,
          <source>SN Computer Science</source>
          <volume>6</volume>
          (
          <year>2025</year>
          )
          <fpage>1</fpage>
          -
          <lpage>25</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>D.</given-names>
            <surname>Muhammad</surname>
          </string-name>
          ,
          <string-name>
            <surname>M.</surname>
          </string-name>
          <article-title>Bendechache, Unveiling the black box: A systematic review of explainable artiifcial intelligence in medical image analysis</article-title>
          ,
          <source>Computational and Structural Biotechnology Journal</source>
          <volume>24</volume>
          (
          <year>2024</year>
          )
          <fpage>542</fpage>
          -
          <lpage>560</lpage>
          . URL: https://www.sciencedirect.com/science/article/pii/S2001037024002642. doi:https://doi.org/10.1016/j.csbj.
          <year>2024</year>
          .
          <volume>08</volume>
          .005.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>M. I.</given-names>
            <surname>Hossain</surname>
          </string-name>
          , G. Zamzmi,
          <string-name>
            <given-names>P. R.</given-names>
            <surname>Mouton</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. S.</given-names>
            <surname>Salekin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Goldgof</surname>
          </string-name>
          ,
          <article-title>Explainable ai for medical data: current methods, limitations, and future directions</article-title>
          ,
          <source>ACM Computing Surveys</source>
          <volume>57</volume>
          (
          <year>2025</year>
          )
          <fpage>1</fpage>
          -
          <lpage>46</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>D.</given-names>
            <surname>Muhammad</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Keles</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Bendechache</surname>
          </string-name>
          ,
          <article-title>Towards explainable deep learning in oncology: Integrating eficientnet-b7 with xai techniques for acute lymphoblastic leukaemia</article-title>
          ,
          <source>in: Proceedings of the 27th European Conference on Artificial Intelligence (ECAI)(Spain</source>
          ,
          <year>2024</year>
          ),
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>R. R.</given-names>
            <surname>Shamir</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Duchin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Kim</surname>
          </string-name>
          , G. Sapiro,
          <string-name>
            <given-names>N.</given-names>
            <surname>Harel</surname>
          </string-name>
          ,
          <article-title>Continuous dice coeficient: a method for evaluating probabilistic segmentations</article-title>
          , arXiv preprint arXiv:
          <year>1906</year>
          .
          <volume>11031</volume>
          (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>J.</given-names>
            <surname>Bertels</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Eelbode</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Berman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Vandermeulen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Maes</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Bisschops</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. B.</given-names>
            <surname>Blaschko</surname>
          </string-name>
          ,
          <article-title>Optimizing the dice score and jaccard index for medical image segmentation: Theory and practice, in: Medical Image Computing and Computer Assisted Intervention-MICCAI</article-title>
          <year>2019</year>
          : 22nd International Conference, Shenzhen, China,
          <source>October 13-17</source>
          ,
          <year>2019</year>
          , Proceedings,
          <source>Part II 22</source>
          , Springer,
          <year>2019</year>
          , pp.
          <fpage>92</fpage>
          -
          <lpage>100</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>D. P.</given-names>
            <surname>Huttenlocher</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G. A.</given-names>
            <surname>Klanderman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W. J.</given-names>
            <surname>Rucklidge</surname>
          </string-name>
          ,
          <article-title>Comparing images using the hausdorf distance</article-title>
          ,
          <source>IEEE Transactions on pattern analysis and machine intelligence</source>
          <volume>15</volume>
          (
          <year>1993</year>
          )
          <fpage>850</fpage>
          -
          <lpage>863</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>W.</given-names>
            <surname>Al-Dhabyani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Gomaa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Khaled</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Fahmy</surname>
          </string-name>
          ,
          <article-title>Dataset of breast ultrasound images</article-title>
          ,
          <source>Data in brief 28</source>
          (
          <year>2020</year>
          )
          <fpage>104863</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>S.</given-names>
            <surname>Zheng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Song</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Leung</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Goodfellow</surname>
          </string-name>
          ,
          <article-title>Improving the robustness of deep neural networks via stability training</article-title>
          ,
          <source>in: Proceedings of the ieee conference on computer vision and pattern recognition</source>
          ,
          <year>2016</year>
          , pp.
          <fpage>4480</fpage>
          -
          <lpage>4488</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>D.</given-names>
            <surname>Muhammad</surname>
          </string-name>
          , I. Ahmad,
          <string-name>
            <given-names>M. I.</given-names>
            <surname>Khalil</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Khalil</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. O.</given-names>
            <surname>Ahmad</surname>
          </string-name>
          ,
          <article-title>A generalized deep learning approach to seismic activity prediction</article-title>
          ,
          <source>Applied Sciences</source>
          <volume>13</volume>
          (
          <year>2023</year>
          ). URL: https://www.mdpi.com/ 2076-3417/13/3/1598. doi:
          <volume>10</volume>
          .3390/app13031598.
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>C.</given-names>
            <surname>Shorten</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T. M.</given-names>
            <surname>Khoshgoftaar</surname>
          </string-name>
          ,
          <article-title>A survey on image data augmentation for deep learning</article-title>
          ,
          <source>Journal of big data 6</source>
          (
          <year>2019</year>
          )
          <fpage>1</fpage>
          -
          <lpage>48</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>D.</given-names>
            <surname>Muhammad</surname>
          </string-name>
          , null Rafiullah,
          <string-name>
            <surname>M.</surname>
          </string-name>
          <article-title>Bendechache, trust: an explainable deep learning framework prediction</article-title>
          ,
          <source>IET Conference Proceedings 2024 digital-library.theiet</source>
          .org/doi/abs/10.1049/icp.
          <year>2024</year>
          .
          <volume>3275</volume>
          .
          <article-title>Improving diagnostic for genitourinary cancer (</article-title>
          <year>2024</year>
          )
          <fpage>47</fpage>
          -
          <lpage>54</lpage>
          . URL: https:// doi:10.1049/icp.
          <year>2024</year>
          .
          <volume>3275</volume>
          . arXiv:https://digital-library.theiet.org/doi/pdf/10.1049/icp.
          <year>2024</year>
          .
          <volume>3275</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <surname>K. O'shea</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Nash</surname>
          </string-name>
          ,
          <article-title>An introduction to convolutional neural networks</article-title>
          ,
          <source>arXiv preprint arXiv:1511.08458</source>
          (
          <year>2015</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>D.</given-names>
            <surname>Muhammad</surname>
          </string-name>
          , I. Ahmed,
          <string-name>
            <given-names>K.</given-names>
            <surname>Naveed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Bendechache</surname>
          </string-name>
          ,
          <article-title>An explainable deep learning approach for stock market trend prediction</article-title>
          ,
          <source>Heliyon</source>
          <volume>10</volume>
          (
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>L. R.</given-names>
            <surname>Dice</surname>
          </string-name>
          ,
          <article-title>Measures of the amount of ecologic association between species</article-title>
          ,
          <source>Ecology</source>
          <volume>26</volume>
          (
          <year>1945</year>
          )
          <fpage>297</fpage>
          -
          <lpage>302</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>T.</given-names>
            <surname>Sorensen</surname>
          </string-name>
          ,
          <article-title>A method of establishing group of equal amplitude in plant sociology based on similarity of species and its application to analyses of the vegetation on danish commons. i kommission hos e, Biologiske Skrifter</article-title>
          ,
          <source>Kongelige Danske Videnskabernes Seleskab</source>
          <volume>5</volume>
          (
          <year>1948</year>
          )
          <fpage>1</fpage>
          -
          <lpage>34</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>T.</given-names>
            <surname>Eelbode</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Bertels</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Berman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Vandermeulen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Maes</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Bisschops</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. B.</given-names>
            <surname>Blaschko</surname>
          </string-name>
          ,
          <article-title>Optimization for medical image segmentation: theory and practice when evaluating with dice score or jaccard index</article-title>
          ,
          <source>IEEE transactions on medical imaging 39</source>
          (
          <year>2020</year>
          )
          <fpage>3679</fpage>
          -
          <lpage>3690</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>