<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>ORCID:</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Semi-Supervised Segmentation of Functional Tissue Units at the Cellular Level</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Volodymyr Sydorskyi</string-name>
          <email>volodymyr.sydorskyi@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Igor Krashenyi</string-name>
          <email>igor.krashenyi@ucu.edu.ua</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Denis Savka</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Oleksandr Zarichkovyi</string-name>
          <email>alexander.zarichkovyi@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Kyiv</institution>
          ,
          <addr-line>03056</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>National Technical University of Ukraine “Igor Sikorsky Kyiv Polytechnic Institute”</institution>
          ,
          <addr-line>37, Prosp. Peremohy</addr-line>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Ukrainian Catholic University</institution>
          ,
          <addr-line>Ilariona Svjentsits'koho St, 17, Lviv, 79000</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
      </contrib-group>
      <volume>000</volume>
      <fpage>0</fpage>
      <lpage>0001</lpage>
      <abstract>
        <p>We present a new method for functional tissue unit segmentation at the cellular level, which utilizes the latest deep learning semantic segmentation approaches together with domain adaptation and semi-supervised learning techniques. This approach allows for minimizing the domain gap, class imbalance, and captures settings influence between HPA and HubMAP datasets. The presented approach achieves comparable with state-of-the-art-result in functional tissue unit segmentation at the semantic segmentation, functional tissue unit, semi-supervised learning</p>
      </abstract>
      <kwd-group>
        <kwd>The</kwd>
        <kwd>source</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        It is estimated that the human body contains approximately 37 trillion cells, and comprehending the
complex relationships and functions among them poses a significant challenge for researchers, requiring
a colossal effort [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. One of the research directions aims to map human body at a cellular level to detect
functional tissue units (FTU). FTU is defined as a unit consisting of a three-dimensional block of cells
centered around a capillary, such that each cell in this block is within diffusion distance from any other
cell in the same block [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. These cellular compositions - cell population neighborhoods are responsible
for performing an organ’s main physiologic functions. Functional tissue units, such as colonic crypts,
renal glomeruli, alveoli, etc. (examples can be observed in Figure 1) have pathobiological relevance
that are essential for modeling and comprehending the development of a disease. However, manually
annotating FTUs is time consuming and costly. At the same time current algorithms suffer from poor
generalizability and low accuracy [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. So the task for the competition was to segment FTU on stained
microscope slides in a way that is invariant to different staining protocols. In this paper a new method
is proposed, which utilizes the latest deep learning semantic segmentation [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] approaches together with
domain adaptation techniques and semi-supervised learning techniques.
      </p>
    </sec>
    <sec id="sec-2">
      <title>2. Related work</title>
      <p>
        One of the most common approaches to functional tissue units segmentation, specifically kidney
glomerulus and colon crypt [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] segmentation is based on the use of supervised learning techniques and
were introduced in the previous Kaggle competition [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. In these methods, the training data consists of
annotated images, where each pixel is labeled as belonging to a particular cell or background. These
techniques typically require a large amount of labeled data to achieve high accuracy, which can be
timeconsuming and expensive to obtain.
EMAIL:
(A.
      </p>
      <p>1);</p>
      <p>2);</p>
      <p>2022 Copyright for this paper by its authors.</p>
      <p>
        Most of these models are heavily inspired by the U-Net [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], UnetPlusPlus [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], FPN architectures [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ],
and DeepLabV3+ [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] in a combination with ImageNet pre-trained backbones such as resnet50_32x4d,
resnet101_32x4d and RegNet [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]. Models used a combination of general data augmentation
techniques such as flipping, rotation, scale shifting, artificial blurring, CutMix [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] and MixUp [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] to
improve model performance. Models were trained using binary cross-entropy and Lovász Hinge loss
[
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] functions, RAdam [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ], Lookahead [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ], AdamW [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ], SGD [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ], and Adam [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ] optimizers.
These models used a dynamic sampling approach to sample tiles of size 512x512, 768x768 and
1024x1024 pixels from regions with visible glomeruli based on the annotations.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Dataset</title>
      <p>
        The dataset includes biopsy slides from several organs, namely kidney, prostate, large intestine,
spleen, and lung. The key feature of the proposed dataset is that it consists of images from two data
sources: HPA [
        <xref ref-type="bibr" rid="ref19 ref20 ref21 ref22 ref23 ref24 ref25 ref26">19-26</xref>
        ] and HuBMAP [
        <xref ref-type="bibr" rid="ref27">27</xref>
        ]. Furthermore, the training data includes only HPA samples,
while the test data comprises a mixture of HPA and HuBMAP samples [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. Additionally, only HubMAP
data was used for the final score (private dataset). The images from the HPA and HuBMAP data sources
differ in staining protocol, pixel sizes, and sample thicknesses [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. Figure 2 provides an example that
illustrates the visual differences between HPA and HubMAP images. The whole slide images in the
HPA and HuBMAP data sources were stained using three distinct protocols. HPA samples were stained
with antibodies visualized with 3,3'-diaminobenzidine (DAB [
        <xref ref-type="bibr" rid="ref28">28</xref>
        ]), counterstained with hematoxylin,
whereas HuBMAP samples were stained using either Periodic acid-Schiff (PAS [
        <xref ref-type="bibr" rid="ref29">29</xref>
        ]), hematoxylin and
eosin stains (H&amp;E [
        <xref ref-type="bibr" rid="ref30">30</xref>
        ]). Each of the staining protocols highlights different cellular structures using
colored dyes, and the final stained slide images vary greatly in color, contrast, and overall image
structure, making direct matching of cellular structures between images less straightforward (see Figure
4). Another crucial feature of the proposed dataset is that HubMAP images have different pixel sizes
for different organs, while for HPA, it is constant (see Table 1) [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ].
      </p>
      <p>
        Finally, the images also differ in tissue section thickness. While all HPA images were sliced with a
fixed thickness of 4 µm, the HuBMAP samples have tissue slice thicknesses ranging from 4 µm for the
spleen and up to 10 µm for the kidney [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], adding another layer of complexity. The training dataset
contained 352 samples along with additional metadata, including the dataset label (HPA or HuBMAP),
organ, image height, image width, pixel size, tissue thickness, age (patient age), and sex (patient sex)
[
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. During the testing stage, we had access to all meta information listed in the train dataset except for
age and sex [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. The test data comprised 550 images, of which 45% were for the public dataset and 65%
for the private dataset [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. The counterplot in Figure 3 also illustrates the class imbalance across organs,
as presented in the training data.
      </p>
    </sec>
    <sec id="sec-4">
      <title>4. Metric and Evaluation</title>
      <p>
        For model evaluation Dice coefficient [
        <xref ref-type="bibr" rid="ref31 ref32">31, 32</xref>
        ] was used, which was simply averaged across all
segmentation masks. For model evaluation metrics on three different datasets were used:
1. Out Of Fold predictions, using 5 Cross-Validation folds [
        <xref ref-type="bibr" rid="ref33">33</xref>
        ]. In order to preserve class
imbalance and make metric more robust, stratification by organ was used.
      </p>
      <p>
        2. Results from the public Kaggle test only on the HubMAP part. While the public Kaggle
test set score was computed using both HPA and HuBMAP images, the final private dataset score was
calculated using only HuBMAP data. We thus decided to focus solely on the HuBMAP score by not
predicting masks for HPA images and adjusting the Kaggle public dataset score by the proportion of
HuBMAP images (roughly 72%) [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ].
      </p>
      <p>3. Results from the private Kaggle test set.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Methods</title>
    </sec>
    <sec id="sec-6">
      <title>5.1. Model Architecture</title>
      <p>
        We have used Unet [
        <xref ref-type="bibr" rid="ref34">34</xref>
        ] and Unet++ [
        <xref ref-type="bibr" rid="ref35">35</xref>
        ] architectures with pre-trained EfficientNet B7 [
        <xref ref-type="bibr" rid="ref36">36</xref>
        ] and
Mix Vision Transformer [
        <xref ref-type="bibr" rid="ref37">37</xref>
        ] encoders. In our experiments, Unet++ showed comparable or better
results compared to the pure Unet decoder, and Mix Vision Transformer outperformed EfficientNet B7
encoders on both Cross-validation and on Private Kaggle Dataset. For our final solution, we used a
simple average of predictions from 15 models using EfficientNet B7 and Mix Vision Transformer
encoder along with Unet and Unet++ style decoder which outperformed either of the single models
(Table 5).
      </p>
    </sec>
    <sec id="sec-7">
      <title>5.2. Data Preparation</title>
      <p>In this challenge, competitors were asked to build a solution that can segment FTUs in a way
that is invariant to the staining protocol (HPA or HubMAP). To achieve this goal, organizers provided
competitors with image data for microscope slides stained using HPA protocol and evaluated solutions
on the mixed HPA+HubMAP dataset for Public Leaderboard and on the HubMAP dataset only for the
Private Leaderboard. Therefore, the biggest challenge for this competition was domain adaptation from
the HPA dataset to HubMAP. In order to solve it we had to adapt our training data in 3 ways:
● Pixel size
● Color space difference
● Tissue thickness difference</p>
      <sec id="sec-7-1">
        <title>Adopting pixel size.</title>
        <p>One of the key points was adapting to wildly varying pixel sizes. The image scales ranged from
6.3um/pixel for the prostate to 0.2um/pixel for the large intestine. We tackled this issue by rescaling
our train dataset to the target HuBMAP resolution. However, to increase the model’s receptive field we
applied additional downscalers for larger images and upscalers for smaller images (prostate). It is
important to note that additional downscalers were also used at the inference stage to avoid changing
the train/test pixel size. We used two datasets: one rescaled to HuBMAP scales and another with the
original HPA scales. The latter one was not only important for HPA predictions (absent in the private
LB) but also to provide some additional scaling information to the model. Therefore, we scaled down
images of each organ by N times in order to match HubMAP pixel size and then by M times to upscale
too small images of organs. Values of N and M can be found in Table 2.</p>
      </sec>
      <sec id="sec-7-2">
        <title>Adopting color space.</title>
        <p>
          The color spaces between HPA and HubMAP datasets were also different due to different stain
methods - DAB [
          <xref ref-type="bibr" rid="ref28">28</xref>
          ] for HPA, PAS [
          <xref ref-type="bibr" rid="ref29">29</xref>
          ], and H&amp;E [
          <xref ref-type="bibr" rid="ref30">30</xref>
          ] for HubMAP (see Figure 4). As the competition
required segmentation of FTUs on slides stained using different staining protocols, we decided to make
the neural network invariant to color variations by applying heavy color augmentations such as
histogram matching [
          <xref ref-type="bibr" rid="ref38">38</xref>
          ] to match the color distribution of the training images to that of HuBMAP
dataset (Figure 5).
We also applied hue-value-saturation, contrast, and gamma augmentations. To provide additional
robustness to scale and geometrical differences in FTU shapes, we also applied a range of geometric
augmentations, which included random flips, rotations, scales, shifts, elastic transforms, and more.
Some competition participants chose to apply stain normalization [
          <xref ref-type="bibr" rid="ref39">39</xref>
          ] to cycle color between different
staining protocols. However, in our experiments, we didn’t see any improvement from stain
normalization, probably because regular stain normalization techniques are specialized for one
particular type of stain and don’t work well when applied to images stained with different protocols.
        </p>
        <p>
          We have gathered additional data from the GTEX portal [
          <xref ref-type="bibr" rid="ref40">40</xref>
          ] and a few images from HubMAP to
which we applied histogram matching [
          <xref ref-type="bibr" rid="ref38">38</xref>
          ] of all train data to GTEX and HubMAP images. The results
of histogram matching may be observed in Figure 5. Besides we have used heavy augmentations
Geometric, Color, Distortions, and Scales. The main idea behind the color augmentation was to suggest
to the model that the color is not important and that it had to look for other features.
        </p>
      </sec>
      <sec id="sec-7-3">
        <title>External data.</title>
        <p>
          We have not tried to solve the problem of tissue thickness explicitly but we have decided to
download additional data from different data sources and apply pseudo-labeling. We used data from
GTEX [
          <xref ref-type="bibr" rid="ref40">40</xref>
          ] and HPA [
          <xref ref-type="bibr" rid="ref19 ref20 ref21 ref22 ref23 ref24 ref25 ref26">19-26</xref>
          ] portals to complement the initial training data. The GTEX data was
especially important here because it was stained similarly to HuBMAP [
          <xref ref-type="bibr" rid="ref27">27</xref>
          ] slides with H&amp;E [
          <xref ref-type="bibr" rid="ref30">30</xref>
          ]. From
GTEX we downloaded prostate, large intestine, kidneys, and spleen data for patients with no apparent
pathologies. We ignored lungs from GTEX as we couldn’t figure out how to segment them and neither
manually nor using pseudo labeling. We were progressively adding GTEX images to our pipeline
ending up with around 140 at the end of the competition, though it is worth mentioning that each image
was quite large measuring tens of thousands of pixels in width and height. From the HPA site, we used
a plethora of DAB [
          <xref ref-type="bibr" rid="ref28">28</xref>
          ] stained slides very similar to those provided by organizers. Overall, we have
added between 57-61K of additional HPA images for each organ.
        </p>
        <p>
          We pseudo-labeled both HPA and HuBMAP images with the best ensemble (according to the Cross
Validation Score) available at the time of labeling. We did not select the most confident pseudo labels
but rather sampled the HPA and GTEX datasets at random at training time. The selection process was
inspired by the pseudo-labeling technique proposed in a semi-supervised paper [
          <xref ref-type="bibr" rid="ref41">41</xref>
          ]. We have repeated
the pseudo-labeling procedure twice. Examples of pseudo-labeled images can be observed in Figure 6.
        </p>
      </sec>
      <sec id="sec-7-4">
        <title>Cutmix.</title>
        <p>
          СutMix [
          <xref ref-type="bibr" rid="ref42">42</xref>
          ] augmentation was among the top contributors to our score. We applied it with a
probability of 0.5 and used uniform distribution to sample which part of the original image to replace
with a patch from a different image. The key trick though was to apply CutMix augmentation within a
single class. Examples of CutMixed image in Figure 7.
        </p>
      </sec>
      <sec id="sec-7-5">
        <title>Filtering Lung samples.</title>
        <p>
          FTUs on lungs were by far the most problematic part of the dataset with our baseline model scoring
a mere 0.05 Dice on Cross Validation vs. 0.69 for the next hardest organ to segment - the spleen.
Baseline model (Unet with EfficientNet B5 [
          <xref ref-type="bibr" rid="ref34 ref36">34, 36</xref>
          ]) Dice on different organs can be observed in Table
3.
        </p>
        <p>There were two major problems with lung FTUs (alveoli): first is inconsistent segmentation of the
FTUs between images (Figure 8), and the second is the shortage of well-segmented samples. Alveoli
on lung images were present in a collapsed and inflated form as well as horizontally and vertically
sectioned. The horizontally sectioned inflated alveoli were the most abundant group, while collapsed
and vertically sectioned images contained only 15 samples on our estimate. When used as a part of the
train set, they generated too much noise and we decided to remove these samples from our training
pipeline.</p>
      </sec>
    </sec>
    <sec id="sec-8">
      <title>5.3. Training Process</title>
      <p>
        We have used 512x512 training crops to train CNN models and 1024x1024 crops for Mix Vision
Transformer [
        <xref ref-type="bibr" rid="ref37">37</xref>
        ] models. Non-empty masks sampled with 0.5 probability. For parameter optimization,
we have used Adam optimizer [
        <xref ref-type="bibr" rid="ref43">43</xref>
        ] with an initial learning rate of 0.001. We reduced it in the training
process with the help of the ReduceLROnPlateau [
        <xref ref-type="bibr" rid="ref44">44</xref>
        ] algorithm with patience 3 by 0.5 factor
monitoring validation dice loss. Initially, the constructed pipeline was a multiclass model with 5
channels - one for each organ. However, as only one class of organ FTUs was present at any given
image we reformulated the task as a binary semantic segmentation with a single channel containing all
the masks no matter what organ was present on an image. Such an approach allowed for improved
generalization and better scores across all organs. To improve model robustness we have used a mixture
of four losses: binary Cross-Entropy, Dice Loss, Focal Loss [
        <xref ref-type="bibr" rid="ref45">45</xref>
        ], and Jacquard Loss [
        <xref ref-type="bibr" rid="ref46">46</xref>
        ]. We have also
trained an EfficientNet model with PointRend head [
        <xref ref-type="bibr" rid="ref47">47</xref>
        ] and scaled loss with a factor of 2. While we
didn’t notice a meaningful performance boost from the PointRend alone we think that its main
contribution was in adding diversity to our model ensemble as well as some regularization. We have
used PyTorch [
        <xref ref-type="bibr" rid="ref48">48</xref>
        ] built-in mixed precision training in order to reduce GPU memory consumption
which allowed us to use a batch size of 32 samples on A100 GPUs.
      </p>
    </sec>
    <sec id="sec-9">
      <title>5.4. Inference Process</title>
      <p>
        For each fold of each of our final models we have averaged model parameters of 3 best checkpoints
by validation dice [
        <xref ref-type="bibr" rid="ref49">49</xref>
        ]. For ensembling, we have simply averaged probability masks from each model.
We have also used Test Time Augmentations [
        <xref ref-type="bibr" rid="ref50">50</xref>
        ] with original images and three flips. We have
removed small regions after thresholding to reduce noisy masks. To do so we have used the next
heuristic:

/
&lt; 
ℎ (1)
      </p>
      <p>OrganTresh for different organs was found empirically by testing its effects on Cross Validation
Dice and can be found in Table 4.</p>
      <p>The results of training 5 models using 5 folds on out-of-fold data, public and private datasets are
outlined in Table 5. Experiment ensembles include 5 models from each experiment and metrics from
them are outlined in Table 6. Results of our approach compared to other top 5 best solutions can be
found in Table 7.</p>
    </sec>
    <sec id="sec-10">
      <title>6. Results</title>
    </sec>
    <sec id="sec-11">
      <title>6.1. Final results</title>
      <p>
        ● Mixed Vision models [
        <xref ref-type="bibr" rid="ref37">37</xref>
        ] outperformed CNN models both on Cross-Validation and Private
test data, which can advise that these models perform better in terms of segmentation quality and
domain adaptation.
      </p>
      <p>
        ● Mean ensemble of CNNs and Mixed Vision models [
        <xref ref-type="bibr" rid="ref37">37</xref>
        ] slightly improved results comparing
to solo CNN or Mixed Vision Transformer [
        <xref ref-type="bibr" rid="ref37">37</xref>
        ] approach.
      </p>
    </sec>
    <sec id="sec-12">
      <title>6.2. Ablation Study</title>
      <p>
        In this section we will outline model performance improvements in terms of Dice score [
        <xref ref-type="bibr" rid="ref31 ref32">31, 32</xref>
        ]
when we have introduced changes, described in previous sections - Table 8.
From this table above we can clearly see that:
      </p>
    </sec>
    <sec id="sec-13">
      <title>7. Conclusion</title>
      <p>● Introduced changes improved Out of Fold Dice and Private Dice, which means that overall
model performance increased both on HPA and HubMAP datasets.</p>
      <p>● Introduced changes decreased, mostly eliminated the gap between Out of Fold Dice, and Private
Dice, which means that they have completed the domain adaptation task between HPA and HubMAP
datasets.</p>
      <p>
        Also, each organ dice improved, especially the lung dice was improved more than 10 times - Table
This paper introduced the FTU segmentation training pipeline, which showed near state-of-the-art
performance both on HPA [
        <xref ref-type="bibr" rid="ref19 ref20 ref21 ref22 ref23 ref24 ref25 ref26">19-26</xref>
        ] and HubMAP [
        <xref ref-type="bibr" rid="ref27">27</xref>
        ] datasets, minimizing the domain gap between
them. Proposed methods allowed the adoption of models from the HPA domain to HubMAP, reducing
the difference in the Dice score between test sets on HPA and HubMAP domains. Also, we have
considerably increased our score on the HPA test set. We believe that the proposed methods can be
used both for increasing the performance of semantic segmentation models on one domain and for
adopting these models from one domain to another.
      </p>
    </sec>
    <sec id="sec-14">
      <title>8. Acknowledgements</title>
      <p>First, we would like to thank the Armed Forces of Ukraine, Security Service of Ukraine, Defence
Intelligence of Ukraine, State Emergency Service of Ukraine for providing safety and security to
participate in this great competition, complete this work, and help science, technology not stop and
move forward. Also, we want to thank the Kaggle team, Google team, Genentech, and Indian
University for hosting HuBMAP + HPA - Hacking the Human Body competition, which gave us all the
needed data and materials to build models, test hypotheses, and write this paper.
9. References</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>Kaggle</given-names>
            <surname>: HuBMAP + HPA - Hacking the Human Body</surname>
          </string-name>
          ,
          <year>2022</year>
          . URL: https://www.kaggle.com/competitions/hubmap-organ-segmentation
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <surname>de Bono</surname>
            <given-names>B</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Grenon</surname>
            <given-names>P</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Baldock</surname>
            <given-names>R</given-names>
          </string-name>
          , Hunter P.:
          <article-title>“Functional tissue units and their primary tissue motifs in multi-scale physiology</article-title>
          .”
          <string-name>
            <given-names>J Biomed</given-names>
            <surname>Semantics</surname>
          </string-name>
          .
          <source>2013 Oct</source>
          <volume>8</volume>
          ;
          <issue>4</issue>
          (
          <issue>1</issue>
          ):
          <fpage>22</fpage>
          . doi:
          <volume>10</volume>
          .1186/2041- 1480-4-22.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <surname>Leah</surname>
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Godwin</surname>
          </string-name>
          , Yingnan Ju, Naveksha Sood, Yashvardhan Jain,
          <string-name>
            <surname>Ellen M. Quardokus</surname>
            , Andreas Bueckle, Teri Longacre, Aaron Horning, Yiing Lin,
            <given-names>Edward D.</given-names>
          </string-name>
          <string-name>
            <surname>Esplin</surname>
            , John W. Hickey,
            <given-names>Michael P.</given-names>
          </string-name>
          <string-name>
            <surname>Snyder</surname>
            ,
            <given-names>N. Heath</given-names>
          </string-name>
          <string-name>
            <surname>Patterson</surname>
          </string-name>
          ,
          <string-name>
            <surname>Jeffrey M. Spraggins</surname>
          </string-name>
          , Katy Börner. “
          <article-title>Robust and generalizable segmentation of human functional tissue units</article-title>
          .” bioRxiv
          <year>2021</year>
          .
          <volume>11</volume>
          .09.467810; doi: https://doi.org/10.1101/
          <year>2021</year>
          .11.09.467810
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>Jonathan</given-names>
            <surname>Long</surname>
          </string-name>
          , Evan Shelhamer, Trevor Darrell. “
          <article-title>Fully Convolutional Networks for Semantic Segmentation</article-title>
          .” (
          <year>2015</year>
          )
          <article-title>IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (</article-title>
          <year>2015</year>
          ). doi:
          <volume>10</volume>
          .1109/CVPR.
          <year>2015</year>
          .7298965
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          <article-title>[5] Kaggle: HuBMAP - Hacking the Kidney</article-title>
          . URL: https://www.kaggle.com/c/hubmap-kidneysegmentation.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <surname>Dice</surname>
          </string-name>
          ,
          <string-name>
            <surname>Lee R</surname>
          </string-name>
          .
          <article-title>"Measures of the Amount of Ecologic Association Between Species"</article-title>
          .
          <source>Ecology</source>
          .
          <volume>26</volume>
          (
          <issue>3</issue>
          ):
          <fpage>297</fpage>
          -
          <lpage>302</lpage>
          . (
          <year>1945</year>
          ). doi:
          <volume>10</volume>
          .2307/1932409. JSTOR 1932409
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>Zongwei</given-names>
            <surname>Zhou</surname>
          </string-name>
          , Md Mahfuzur Rahman Siddiquee, Nima Tajbakhsh and
          <string-name>
            <given-names>Jianming</given-names>
            <surname>Liang</surname>
          </string-name>
          . “UNet++
          <article-title>: A Nested U-Net Architecture for Medical Image Segmentation</article-title>
          ” arXiv:
          <year>1807</year>
          .
          <volume>10165</volume>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <surname>Tsung-Yi</surname>
            <given-names>Lin</given-names>
          </string-name>
          , Piotr Dollár, Ross Girshick, Kaiming He, Bharath Hariharan, Serge Belongie: “
          <source>Feature Pyramid Networks for Object Detection” arXiv:1612.03144</source>
          (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <surname>Liang-Chieh</surname>
            <given-names>Chen</given-names>
          </string-name>
          , George Papandreou, Florian Schroff, Hartwig Adam: “
          <source>Rethinking Atrous Convolution for Semantic Image Segmentation” arXiv:1706.05587</source>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>Jing</surname>
            <given-names>Xu</given-names>
          </string-name>
          , Yu Pan, Xinglin Pan, Steven Hoi, Zhang Yi, Zenglin Xu:
          <article-title>“RegNet: Self-Regulated Network for Image Classification</article-title>
          ” arXiv:
          <fpage>2101</fpage>
          .00590 (
          <year>2021</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>Sangdoo</surname>
            <given-names>Yun</given-names>
          </string-name>
          , Dongyoon Han, Seong Joon Oh,
          <string-name>
            <given-names>Sanghyuk</given-names>
            <surname>Chun</surname>
          </string-name>
          , Junsuk Choe, Youngjoon Yoo:
          <article-title>“CutMix: Regularization Strategy to Train Strong Classifiers with Localizable Features” arXiv:</article-title>
          <year>1905</year>
          .
          <volume>04899</volume>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <surname>Hongyi</surname>
            <given-names>Zhang</given-names>
          </string-name>
          , Moustapha Cisse,
          <string-name>
            <given-names>Yann N.</given-names>
            <surname>Dauphin</surname>
          </string-name>
          , David Lopez-Paz:
          <article-title>“mixup: Beyond Empirical Risk Minimization</article-title>
          ” arXiv:
          <fpage>1710</fpage>
          .09412 (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <surname>Jiaqian</surname>
            <given-names>Yu</given-names>
          </string-name>
          , Matthew Blaschko: “
          <article-title>The Lovász Hinge: A Novel Convex Surrogate for Submodular Losses</article-title>
          ” arXiv:
          <fpage>1512</fpage>
          .07797 (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <surname>Liyuan</surname>
            <given-names>Liu</given-names>
          </string-name>
          , Haoming Jiang, Pengcheng He, Weizhu Chen, Xiaodong Liu, Jianfeng Gao, Jiawei Han: “
          <article-title>On the Variance of the Adaptive Learning Rate</article-title>
          and Beyond” arXiv:
          <year>1908</year>
          .
          <volume>03265</volume>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <surname>Michael</surname>
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Zhang</surname>
          </string-name>
          , James Lucas, Geoffrey Hinton, Jimmy Ba:
          <article-title>“Lookahead Optimizer: k steps forward, 1 step back</article-title>
          ” arXiv:
          <year>1907</year>
          .
          <volume>08610</volume>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <surname>Ilya</surname>
            <given-names>Loshchilov</given-names>
          </string-name>
          , Frank Hutter: “
          <source>Decoupled Weight Decay Regularization” arXiv:1711.05101</source>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <surname>Herbert</surname>
            <given-names>Robbins</given-names>
          </string-name>
          , Sutton Monro: “
          <string-name>
            <given-names>A Stochastic</given-names>
            <surname>Approximation</surname>
          </string-name>
          <string-name>
            <surname>Method</surname>
          </string-name>
          ” Ann. Math. Statist.
          <volume>22</volume>
          (
          <issue>3</issue>
          ):
          <fpage>400</fpage>
          -
          <lpage>407</lpage>
          (
          <year>1951</year>
          ). DOI:
          <volume>10</volume>
          .1214/aoms/1177729586
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <surname>Diederik</surname>
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Kingma</surname>
          </string-name>
          , Jimmy Ba:
          <article-title>“Adam: A Method for Stochastic Optimization</article-title>
          ” arXiv:
          <fpage>1412</fpage>
          .6980 (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <surname>Uhlén</surname>
            <given-names>M</given-names>
          </string-name>
          et al., “
          <article-title>Tissue-based map of the human proteome</article-title>
          .”
          <source>Science</source>
          (
          <year>2015</year>
          ) PubMed: 25613900 doi: 10.1126/science.1260419
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <surname>Thul</surname>
            <given-names>PJ</given-names>
          </string-name>
          et al.,
          <article-title>“A subcellular map of the human proteome</article-title>
          .” Science. (
          <year>2017</year>
          ) PubMed: 28495876 doi: 10.1126/science.aal3321
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <surname>Sjöstedt</surname>
            <given-names>E</given-names>
          </string-name>
          et al.,
          <article-title>“An atlas of the protein-coding genes in the human, pig, and mouse brain</article-title>
          .”
          <source>Science</source>
          . (
          <year>2020</year>
          ) PubMed: 32139519 doi: 10.1126/science.aay5947
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <surname>Karlsson</surname>
            <given-names>M</given-names>
          </string-name>
          et al.,
          <article-title>“A single-cell type transcriptomics map of human tissues</article-title>
          .
          <source>” Sci Adv.</source>
          (
          <year>2021</year>
          ) PubMed: 34321199 doi: 10.1126/sciadv.abh2169
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <surname>Uhlen</surname>
            <given-names>M</given-names>
          </string-name>
          et al.,
          <article-title>“A pathology atlas of the human cancer transcriptome</article-title>
          .
          <source>” Science</source>
          . (
          <year>2017</year>
          ) PubMed: 28818916 doi: 10.1126/science.aan2507
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <surname>Uhlen</surname>
            <given-names>M</given-names>
          </string-name>
          et al.,
          <article-title>“A genome-wide transcriptomic analysis of protein-coding genes in human blood cells</article-title>
          .”
          <source>Science</source>
          . (
          <year>2019</year>
          ) PubMed: 31857451 doi: 10.1126/science.aax9198
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <surname>Sjöstedt</surname>
            <given-names>E</given-names>
          </string-name>
          et al.,
          <article-title>“An atlas of the protein-coding genes in the human, pig, and mouse brain</article-title>
          .”
          <source>Science</source>
          . (
          <year>2020</year>
          ) PubMed: 32139519 doi: 10.1126/science.aay5947
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [26]
          <string-name>
            <surname>Uhlén</surname>
            <given-names>M</given-names>
          </string-name>
          et al., “
          <article-title>The human secretome</article-title>
          .
          <source>” Sci Signal.</source>
          (
          <year>2019</year>
          ) PubMed: 31772123 doi: 10.1126/scisignal.aaz0274
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [27]
          <string-name>
            <surname>Snyder</surname>
            ,
            <given-names>M.P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lin</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Posgai</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          et al. “
          <article-title>The human body at cellular resolution: the NIH Human Biomolecular Atlas Program</article-title>
          .
          <source>” Nature</source>
          <volume>574</volume>
          ,
          <fpage>187</fpage>
          -
          <lpage>192</lpage>
          (
          <year>2019</year>
          ). doi:
          <volume>10</volume>
          .1038/s41586-019-1629-x.
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          [28]
          <string-name>
            <surname>Litwin</surname>
            <given-names>JA.</given-names>
          </string-name>
          <article-title>Histochemistry and cytochemistry of 3,3'-diaminobenzidine. A review</article-title>
          .
          <source>Folia Histochem Cytochem (Krakow)</source>
          .
          <year>1979</year>
          ;
          <volume>17</volume>
          (
          <issue>1</issue>
          ):
          <fpage>3</fpage>
          -
          <lpage>28</lpage>
          . PMID:
          <volume>220157</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          [29]
          <string-name>
            <surname>GOMORI G.</surname>
          </string-name>
          <article-title>The periodic-acid Schiff stain</article-title>
          .
          <source>Am J Clin Pathol</source>
          .
          <year>1952</year>
          Mar;
          <volume>22</volume>
          (
          <issue>3</issue>
          ):
          <fpage>277</fpage>
          -
          <lpage>81</lpage>
          . doi:
          <volume>10</volume>
          .1093/ajcp/22.3_ts.277. PMID:
          <volume>14902736</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          [30]
          <string-name>
            <surname>Feldman</surname>
            <given-names>AT</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wolfe</surname>
            <given-names>D.</given-names>
          </string-name>
          <article-title>Tissue processing and hematoxylin and eosin staining</article-title>
          .
          <source>Methods Mol Biol</source>
          .
          <year>2014</year>
          ;
          <volume>1180</volume>
          :
          <fpage>31</fpage>
          -
          <lpage>43</lpage>
          . doi:
          <volume>10</volume>
          .1007/978-1-
          <fpage>4939</fpage>
          -1050-
          <issue>2</issue>
          _
          <fpage>3</fpage>
          . PMID:
          <volume>25015141</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          [31]
          <string-name>
            <surname>Sørensen</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          "
          <article-title>A method of establishing groups of equal amplitude in plant sociology based on similarity of species and its application to analyses of the vegetation on Danish commons"</article-title>
          .
          <source>Kongelige Danske Videnskabernes Selskab</source>
          .
          <volume>5</volume>
          (
          <issue>4</issue>
          ):
          <fpage>1</fpage>
          -
          <lpage>34</lpage>
          . (
          <year>1948</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          [32]
          <string-name>
            <surname>Dice</surname>
          </string-name>
          ,
          <string-name>
            <surname>Lee R</surname>
          </string-name>
          .
          <article-title>"Measures of the Amount of Ecologic Association Between Species"</article-title>
          .
          <source>Ecology</source>
          .
          <volume>26</volume>
          (
          <issue>3</issue>
          ):
          <fpage>297</fpage>
          -
          <lpage>302</lpage>
          . (
          <year>1945</year>
          ). doi:
          <volume>10</volume>
          .2307/1932409. JSTOR 1932409
        </mixed-citation>
      </ref>
      <ref id="ref33">
        <mixed-citation>
          [33]
          <string-name>
            <surname>Allen</surname>
            ,
            <given-names>David M. "</given-names>
          </string-name>
          <article-title>The Relationship between Variable Selection and Data Agumentation and a Method for Prediction"</article-title>
          .
          <source>Technometrics</source>
          .
          <volume>16</volume>
          (
          <issue>1</issue>
          ):
          <fpage>125</fpage>
          -
          <lpage>127</lpage>
          . (
          <year>1974</year>
          ). doi:
          <volume>10</volume>
          .2307/1267500. JSTOR 1267500
        </mixed-citation>
      </ref>
      <ref id="ref34">
        <mixed-citation>
          [34]
          <string-name>
            <surname>Olaf</surname>
            <given-names>Ronneberger</given-names>
          </string-name>
          , Philipp Fischer and
          <string-name>
            <given-names>Thomas</given-names>
            <surname>Brox</surname>
          </string-name>
          . “U-Net:
          <article-title>Convolutional Networks for Biomedical Image Segmentation.” Medical Image Computing</article-title>
          and
          <string-name>
            <surname>Computer-Assisted</surname>
            <given-names>Intervention - MICCAI</given-names>
          </string-name>
          <year>2015</year>
          .
          <source>MICCAI 2015. Lecture Notes in Computer Science()</source>
          , vol
          <volume>9351</volume>
          . Springer, Cham. doi:
          <volume>10</volume>
          .1007/978-3-
          <fpage>319</fpage>
          -24574-4_
          <fpage>28</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref35">
        <mixed-citation>
          [35]
          <string-name>
            <surname>Zongwei</surname>
            <given-names>Zhou</given-names>
          </string-name>
          , Md Mahfuzur Rahman Siddiquee, Nima Tajbakhsh, Jianming Liang. “UNet++
          <article-title>: A Nested U-Net Architecture for Medical Image Segmentation.” Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support</article-title>
          .
          <source>DLMIA ML-CDS 2018 2018. Lecture Notes in Computer Science()</source>
          , vol
          <volume>11045</volume>
          . Springer, Cham. doi:
          <volume>10</volume>
          .1007/978-3-
          <fpage>030</fpage>
          -00889-
          <issue>5</issue>
          _
          <fpage>1</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref36">
        <mixed-citation>
          [36]
          <string-name>
            <surname>Mingxing</surname>
            ,
            <given-names>Tan.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Quoc</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Le</surname>
          </string-name>
          .
          <article-title>"EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks." arXiv:</article-title>
          <year>1905</year>
          .
          <volume>11946</volume>
          (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref37">
        <mixed-citation>
          [37]
          <string-name>
            <surname>Jie-Neng</surname>
            <given-names>Chen</given-names>
          </string-name>
          , Shuyang Sun, Ju He, Philip Torr, Alan Yuille and
          <string-name>
            <given-names>Song</given-names>
            <surname>Bai</surname>
          </string-name>
          . “TransMix:
          <article-title>Attend to Mix for Vision Transformers</article-title>
          .” arXiv:
          <fpage>2111</fpage>
          .09833 (
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref38">
        <mixed-citation>
          [38]
          <article-title>Scikit-image Histogram matching URL: https://scikitimage</article-title>
          .org/docs/dev/auto_examples/color_exposure/plot_histogram_matching.html
        </mixed-citation>
      </ref>
      <ref id="ref39">
        <mixed-citation>
          [39]
          <string-name>
            <given-names>Marc</given-names>
            <surname>Macenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Marc</given-names>
            <surname>Niethammer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. S.</given-names>
            <surname>Marron</surname>
          </string-name>
          , David Borland, John T. Woosley, Xiaojun Guan, Charles Schmitt, and Nancy E.
          <article-title>Thomas “A METHOD FOR NORMALIZING HISTOLOGY SLIDES FOR QUANTITATIVE ANALYSIS</article-title>
          ” https://www.cs.unc.edu/~mn/sites/default/files/macenko2009.pdf
        </mixed-citation>
      </ref>
      <ref id="ref40">
        <mixed-citation>
          [40]
          <string-name>
            <given-names>GTEx</given-names>
            <surname>Histology Viewer</surname>
          </string-name>
          <string-name>
            <surname>URL</surname>
          </string-name>
          :https://www.gtexportal.org/home/histologyPage
        </mixed-citation>
      </ref>
      <ref id="ref41">
        <mixed-citation>
          [41]
          <string-name>
            <surname>Yauhen</surname>
            <given-names>Babakhin</given-names>
          </string-name>
          , Artsiom Sanakoyeu, Hirotoshi Kitamura. “
          <article-title>Semi-Supervised Segmentation of Salt Bodies in Seismic Images using an Ensemble of Convolutional Neural Networks</article-title>
          .” arXiv:
          <year>1904</year>
          .
          <volume>04445</volume>
          (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref42">
        <mixed-citation>
          [42]
          <string-name>
            <surname>Sangdoo</surname>
            <given-names>Yun</given-names>
          </string-name>
          , Dongyoon Han, Seong Joon Oh,
          <string-name>
            <given-names>Sanghyuk</given-names>
            <surname>Chun</surname>
          </string-name>
          , Junsuk Choe and
          <string-name>
            <given-names>Youngjoon</given-names>
            <surname>Yoo</surname>
          </string-name>
          . “
          <article-title>CutMix: Regularization Strategy to Train Strong Classifiers with Localizable Features</article-title>
          .” in 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Korea (South),
          <year>2019</year>
          pp.
          <fpage>6022</fpage>
          -
          <lpage>6031</lpage>
          . doi:
          <volume>10</volume>
          .1109/ICCV.
          <year>2019</year>
          .00612
        </mixed-citation>
      </ref>
      <ref id="ref43">
        <mixed-citation>
          [43]
          <string-name>
            <surname>Kingma</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Ba</surname>
          </string-name>
          , J. “
          <article-title>Adam: A Method for Stochastic Optimization</article-title>
          .
          <source>” Proceedings of the 3rd International Conference on Learning Representations (ICLR</source>
          <year>2015</year>
          ).
          <article-title>(</article-title>
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref44">
        <mixed-citation>
          [44] PyTorch: ReduceLROnPlateau,
          <year>2022</year>
          . URL: https://pytorch.org/docs/stable/generated/torch.optim.lr_scheduler.ReduceLROnPlateau.html
        </mixed-citation>
      </ref>
      <ref id="ref45">
        <mixed-citation>
          [45]
          <string-name>
            <surname>T. -Y. Lin</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Goyal</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Girshick</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>He</surname>
            and
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Dollár</surname>
          </string-name>
          ,
          <article-title>"Focal Loss for Dense Object Detection."</article-title>
          <source>IEEE International Conference on Computer Vision</source>
          (ICCV),
          <year>2017</year>
          , pp.
          <fpage>2999</fpage>
          -
          <lpage>3007</lpage>
          . (
          <year>2017</year>
          ) doi: 10.1109/ICCV.
          <year>2017</year>
          .
          <volume>324</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref46">
        <mixed-citation>
          [46]
          <string-name>
            <surname>Jaccard</surname>
            ,
            <given-names>Paul. "</given-names>
          </string-name>
          <article-title>The Distribution of the Flora in the Alpine Zone.1"</article-title>
          . New Phytologist.
          <volume>11</volume>
          (
          <issue>2</issue>
          ):
          <fpage>37</fpage>
          -
          <lpage>50</lpage>
          . (
          <year>1912</year>
          ). doi:
          <volume>10</volume>
          .1111/j.1469-
          <fpage>8137</fpage>
          .
          <year>1912</year>
          .tb05611.
          <source>x. ISSN 0028-646X</source>
        </mixed-citation>
      </ref>
      <ref id="ref47">
        <mixed-citation>
          [47]
          <string-name>
            <surname>Alexander</surname>
            <given-names>Kirillov</given-names>
          </string-name>
          , Yuxin Wu, Kaiming He and
          <string-name>
            <given-names>Ross</given-names>
            <surname>Girshick</surname>
          </string-name>
          . “
          <article-title>PointRend: Image Segmentation as Rendering”</article-title>
          .
          <source>CVPR</source>
          <year>2020</year>
          :
          <fpage>9796</fpage>
          -
          <lpage>9805</lpage>
          (
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref48">
        <mixed-citation>
          [48]
          <string-name>
            <surname>Paszke</surname>
            <given-names>A</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gross</surname>
            <given-names>S</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Massa</surname>
            <given-names>F</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lerer</surname>
            <given-names>A</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bradbury</surname>
            <given-names>J</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chanan</surname>
            <given-names>G</given-names>
          </string-name>
          , et al. “
          <article-title>PyTorch: An Imperative Style, High-Performance Deep Learning Library</article-title>
          .”
          <source>In: Advances in Neural Information Processing Systems</source>
          <volume>32</volume>
          . Curran Associates, Inc.; pp.
          <fpage>8024</fpage>
          -
          <lpage>35</lpage>
          . (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref49">
        <mixed-citation>
          [49]
          <string-name>
            <surname>Izmailov</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Podoprikhin</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Garipov</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vetrov</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          , &amp; Wilson,
          <string-name>
            <surname>A. G.</surname>
          </string-name>
          (
          <year>2018</year>
          ).
          <article-title>Averaging weights leads to wider optima and better generalization</article-title>
          . In R. Silva,
          <string-name>
            <given-names>A.</given-names>
            <surname>Globerson</surname>
          </string-name>
          , &amp; A.
          <string-name>
            <surname>Globerson</surname>
          </string-name>
          (Eds.),
          <source>34th Conference on Uncertainty in Artificial Intelligence</source>
          <year>2018</year>
          ,
          <string-name>
            <surname>UAI</surname>
          </string-name>
          <year>2018</year>
          (pp.
          <fpage>876</fpage>
          -
          <lpage>885</lpage>
          ).
          <source>(34th Conference on Uncertainty in Artificial Intelligence</source>
          <year>2018</year>
          , UAI 2018; Vol.
          <volume>2</volume>
          ).
          <source>Association For Uncertainty in Artificial Intelligence (AUAI).</source>
        </mixed-citation>
      </ref>
      <ref id="ref50">
        <mixed-citation>
          [50]
          <string-name>
            <surname>Shorten</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Khoshgoftaar</surname>
            ,
            <given-names>T.M. “</given-names>
          </string-name>
          <article-title>A survey on Image Data Augmentation for Deep Learning</article-title>
          .
          <source>” J Big Data</source>
          <volume>6</volume>
          ,
          <issue>60</issue>
          (
          <year>2019</year>
          ).
          <source>doi: 10.1186/s40537-019-0197-0.</source>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>