<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>An Embedded Recurrent Neural Network-based Model for Endoscopic Semantic Segmentation</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Mahmood Haithami</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Amr Ahmed</string-name>
          <email>Amr.Ahmed@nottingham.edu.my</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Iman Yi Liao</string-name>
          <email>Iman.Liao@nottingham.edu.my</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Hamid Jalab</string-name>
          <email>hamidjalab@um.edu.my</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Computer Science Department,University of Nottingham Malaysia Campus</institution>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Computer System and Technology Department, University of Malaya</institution>
        </aff>
      </contrib-group>
      <abstract>
        <p>Detecting cancers at their early stage would decrease mortality rate. For instance, detecting all polyps during colonoscopy would increase the chances of a better prognoses. However, endoscopists are facing dificulties due to the heavy workload of analyzing endoscopic images. Hence, assisting endoscopist while screening would decrease polyp miss rate. In this study, we propose a new deep learning segmentation model to segment polyps found in endoscopic images extracted during Colonoscopy screening. The propose model modifies SegNet architecture to embed Gated recurrent units (GRU) units within the convolution layers to collect contextual information. Therefore, both global and local information are extracted and propagated through the entire layers. This has led to better segmentation performance compared to that of using state of the art SegNet. Four experiments were conducted and the proposed model achieved a better intersection over union “IoU” by 1.36%, 1.71%, and 1.47% on validation sets and 0.24% on a test set, compared to the state of the art SegNet.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;SegNet</kwd>
        <kwd>GRU</kwd>
        <kwd>Embedded RNN</kwd>
        <kwd>Polyp Segmentation</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        According to National Institute of Diabetes and Digestive and Kidney Diseases [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], 60 to 70
million people are afected by a gastrointestinal disease. Malignancies such as esophageal and
colorectal cancer are in an increasing rate in western countries [
        <xref ref-type="bibr" rid="ref2 ref3 ref4">2, 3, 4</xref>
        ]. Early detection and
removal of such malignant tissues using endoscopy would reduce the risk of developing a cancer.
However, endoscopists are facing dificulties due to the heavy workload of analyzing endoscopic
images [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], subtle lesions, or lack of experience [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. Therefore, researchers have been proposing
deep learning models to help endoscopists marking malignancies during the screening [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ].
      </p>
      <p>
        Polyp segmentation is considered to be a challenging task due to the non-uniformity of the
gastrointestinal tract. Hence, researchers tend to modify on-the-shelf deep learning models
that proved to be eficient in a specific domain [
        <xref ref-type="bibr" rid="ref8 ref9">8, 9</xref>
        ]. Also, the lack of big and representative
datasets in the endoscopy domain is a persistent challenge which clamp the performance of
deep learning models. Therefore, transfer learning is frequently used to enhance a model’s
capability to segment diseases [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] and artifacts [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] found in endoscopic images. In other cases,
augmentation is adopted instead as in [
        <xref ref-type="bibr" rid="ref12 ref13">12, 13</xref>
        ].
      </p>
      <p>
        However it is noticed that Recurrent Neural Network (RNN) models are not studied enough in
the literature. According to a recent survey [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ], only 4 published studies employed RNN models
to detect and segment colorectal polyps. Therefore in this research, we attempted to study the
feasibility of using RNN. Gated recurrent unit (GRU) [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ] - a type of RNN model- is selected to
enhance the capabilities of SegNet [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ] model to segment polyps. The proposed model contains
four GRU units which are embedded into SegNet [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ] to add contextual information to the
feature maps. We conducted four experiments using EndoCV21 [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ] dataset which was posted
for “3rd International Endoscopy Computer Vision challenge and workshop” challenge [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ].
The proposed model achieved intersection over union ”IoU” better than the state-of-the-art
SegNet [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ] by a 1.51% on average for the validation sets and 0.24% on unseen test set.
      </p>
    </sec>
    <sec id="sec-2">
      <title>2. Methodology</title>
      <p>In this section, the used dataset as well as the proposed model would be discussed and elaborated
in a detailed fashion.</p>
      <sec id="sec-2-1">
        <title>2.1. Dataset</title>
        <p>
          The used dataset is EndoCV21 which was created using diferent datacenters [
          <xref ref-type="bibr" rid="ref17">17</xref>
          ]. The database
can be used for polyp segmentation as well as detection since both the mask and bounding box
annotations in VOC format were provided. An interesting fact about this database is that it is
divided into five sets, each corresponding to a medical center. Center-wise split is provided
to encourage researchers to develop generalizable methods [
          <xref ref-type="bibr" rid="ref17 ref18">17, 18</xref>
          ]. The five annotated single
frame datasets are Data_C1, Data_C2, Data_C3, Data_C4, and Data_C5 containing 256, 305, 457,
227, and 207 images, respectively. The images may have more than one polyp or no polyp at all.
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Proposed model</title>
        <p>
          Semantic segmentation requires integration and balancing of local information as well as
global information at various spatial scales [19]. Conventional CNN models have some degree
of spatial invariance due to pooling, though, they neglect global context information [19].
Therefore, diferent techniques are proposed in the literature to make CNN aware of contextual
information such as applying post-processing refinement, dilated convolution, multi-scale
convolution\aggregation, and applying sequence modeling “RNN”. Inspired by [20], we propose
a segmentation model that employ Gated Recurrent Unit (GRU) [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ] within the layers of SegNet
[
          <xref ref-type="bibr" rid="ref16">16</xref>
          ].
        </p>
        <p>As opposed to [20], our proposed model embeds a GRU unit in each convolution stage within
SegNet as shown in Figure2. A GRU unit is employed before the last convolution layer within
each convolution stage. The total number of stages is five for the encoder as well as the decoder
as depicted in Figure2. The GRU units are used to produce relevant global information [20].
The output feature map from the GRU units is then concatenated with the input feature map
and passed to the next stage. By doing so the local features at each pixel position with respect
to the whole input feature map would be implicitly encoded [20]. At each stage, only one
GRU unit sweeps over the obtained feature map horizontally then vertically as depicted in
Figure 2. The size of GRU hidden state is half of the channel size of the input feature map.
Hence concatenating the GRU’s output feature map after sweeping it in two directions will
produce an output feature map with a size equal to the input feature map, as depicted in Figure 2.
The encoder as well as the decoder share the same GRU units for each corresponding stage as
illustrated in Figure 2.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Experimental Results</title>
      <p>The used Hyperparameters as will as the used metrics will be explained. Then the experimental
results will be presented.</p>
      <sec id="sec-3-1">
        <title>3.1. Hyperparameters</title>
        <p>
          Both the proposed model and SegNet [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ] had the same hyperparameters. Pytorch and
Torchvision were used to conduct all experiments. The learning rate lr=0.005 and the batch size was 4
images. Adam optimization method was employed, and weighted Cross Entropy loss was used.
The loss was weighted to mitigate the efect of imbalanced classes (i.e., class 0 “healthy” and
class 1 “polyps”). Learning rate decay was employed with Multiplicative factor “gamma=0.8”.
Training and validation percentages are 80% and 20%, respectively.
        </p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Metrics</title>
        <p>Three metrics were used in the experiments, namely, pixel accuracy and Intersection over Union
“IoU”. The pixel accuracy is defined as follows:
 =</p>
        <p>+  
  +   +   +  
where TP and TN are the true positive “polyp” and true negative “background”, respectively.
While FP and FN are false positive and false negative, respectively. As opposed to pixel accuracy
metrics, only TP “polyp” is considered in calculating the IoU:
Finally, the Dice similarity is calculated based on IoU as follows:
 =</p>
        <p>+   +  
 =</p>
        <p>
          2
 + 1
EndoCV2021 leaderboard also included the generalisation metrics similar to detection
generalisation defined in [
          <xref ref-type="bibr" rid="ref18">18</xref>
          ] for assessing performance gaps.
        </p>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. Results</title>
        <p>It is important to mention that only polyp class is considered for calculating the IoU. Adding a
background class “healthy” in the IoU’s calculation would give us exaggerated results since the
images’ pixels are biased toward the background. The images as well as the masks are resized
for computational eficiency reasons. One experiment was conducted with a resize factor of 3
and the rest were conducted with a resize factor of 6 (i.e., H=1080/6, W=1380/6).</p>
        <p>
          Four experiments in total were conducted to assess the performance of the proposed model
against the state-of-the-art SegNet [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ]. For the first experiment and the second experiment,
Data_C1, Data_C2, and Data_C3 datasets were used with an image resize factor of 3 and 6,
respectively. The aim is to see whether resizing the images would afect the performance of the
proposed model in terms IoU metric. The total number of single frame images in this dataset is
1452. Figure 3 depicts metric curves (i.e., loss, IoU, and pixels accuracy) for the training and
validation across 50 epochs. The third experiment employed diferent dataset in which Data_C1,
Data_C4, and Data_C5 are used for training and validation. The images are resized by a factor
of 6 and the total number of images is 690. The obtained highest IoU for the training and the
validation set are summarized in Table 1 and Table 2, respectively. The mean and the standard
deviation are calculated across the three experiments for the validation set. Dice similarity are
calculated for each experiment based on the obtained IoU.
        </p>
        <p>The fourth experiment was conducted to test the generalizability of the proposed model to
correctly segment the polyp with contrast to SegNet as depicted in Table3. The best models’
weights are selected based on the IoU obtained from the second experiment. The test set
is composed of Data_C4 and Data_C5 with a total of 434 images. Note that in the second
experiment we didn’t use Data_C4 and Data_C5.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Discussion and Conclusion</title>
      <sec id="sec-4-1">
        <title>4.1. Discussion</title>
        <p>Experiment1: From Figure 3, it is clear that the proposed model is able to learn from the
training data better than SegNet. Specifically speaking, the training IoU graph of the proposed
model is higher than of the one obtained by SegNet. Furthermore, the highest IoU on the
validation set is 0.1837 and 0.1974 for SegNet and the proposed model, respectively. As it can
be seen from Table 1, the recorded highest IoU during the training were 0.2082 and 0.3033 for
SegNet and the proposed model, respectively. This is an indication that the embedded GRU
units enhanced the model’s segmentation capabilities.</p>
        <p>Experiment2 and Experiment3: As opposed to Experiment1, the gap between the achieved
highest IoU during the training of the two models is less, as it can be observed from Table 1.
Note that in Experiment1 the images are resized to a factor of 3 while in Experiment2 and
Experiment3 the image are resize by a factor of 6. This observation is an indication that
SegNet learns contextual information better with smaller images resolution given the fact that
SegNet backbone is based on VGG16 which has only 16 convolutional layers. Reducing the
spatial dimensions of input images will enhance the performance of the convolution layers
of VGG11 which in turn will enhance the overall segmentation’s performance. Furthermore,
using max-pooling layers will create denser feature maps if small images were used instead of
larger resolution images. Nonetheless, still the proposed model achieved the highest IoU on
the validation set as well with a diference of 1.72%, 1.47% in Experiment2 and Experiment 3,
respectively. The IoU results for the training and validation sets are summarized in Table 1 and
Table 2.</p>
        <p>Experiment4 (test set): The proposed model achieved better IoU on the test set than SegNet
with IoU=0.1063 for the former and IoU=0.1039 for the later, as seen in Table 3. However,
compared to the other experiments, the diference in IoU between SegNet and the proposed
model is relatively small (i.e., the diference is 0.24%). Although the embedded GRU units
enhanced the model’s ability to segment polyps found in images, they did not considerably
help the model, relative to the state of the art SegNet, to generalize the learned segmentation
capabilities to new unseen images. Nevertheless, it is clear from the experimental results that
the proposed embedding GRU units enhanced the model’s ability to learn contextual features,
which in turn enhanced the overall segmentation performance. In general, the proposed model
achieved IoU better than the state-of-the-art SegNet in all experiments.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Conclusion</title>
        <p>
          The conclusions can be summarized as follows:
• We proposed a new segmentation model based on SegNet [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ] and GRU [] RNN network
for colorectal polyp segmentation. Four GRU units are embedded within the SegNet
convolution layers and shared between the encoder and decoder.
• The IoU metric demonstrated that the segmentation capabilities are better than the
stateof-the-art SegNet. The reason of this enhancement is the ability of GRU units to extract
global information within the feature maps.
• The proposed model achieved better IoU than the state-of-the-art SegNet. The mean
IoU on the validation sets for the proposed mode and SegNet are 0.1897 and 0.1746,
respectively. While in the test set the proposed model and SegNet achieved 0.1063 and
0.1039, respectively.
• The diference in IoU on the test set between the proposed model and SegNet is relatively
small even though the proposed model showed better IoU on the validation sets. Therefore,
the generalizability property of the embedded GRU units needs to be studied further with
diferent embedding techniques.
        </p>
        <p>Soberanis-Mukul, S. Albarqouni, X. Wang, C. Wang, S. Watanabe, I. Oksuz, Q. Ning, S. Yang,
M. A. Khan, X. W. Gao, S. Realdon, M. Loshchenov, J. A. Schnabel, J. E. East, G. Wagnieres,
V. B. Loschenov, E. Grisan, C. Daul, W. Blondel, J. Rittscher, An objective comparison
of detection and segmentation algorithms for artefacts in clinical endoscopy, Scientific
Reports 10 (2020) 2748. doi:10.1038/s41598-020-59413-5.
[19] A. Garcia-Garcia, S. Orts-Escolano, S. Oprea, V. Villena-Martinez, P. Martinez-Gonzalez,
J. Garcia-Rodriguez, A survey on deep learning techniques for image and video semantic
segmentation, Applied Soft Computing Journal 70 (2018) 41–65.
[20] F. Visin, M. Ciccone, A. Romero, K. Kastner, K. Cho, Y. Bengio, M. Matteucci, A. Courville,
Reseg: A recurrent neural network-based model for semantic segmentation, 2016 IEEE
Conference on Computer Vision and Pattern Recognition Workshops (CVPRW) (2015)
426–433.</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <article-title>[1] Digestive diseases statistics for the united states | niddk, 2020</article-title>
          . URL: https://www.niddk.nih. gov/health-information/health-statistics/digestive-diseases#all, [Online; accessed 2020-
          <volume>02</volume>
          - 12].
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>A.</given-names>
            <surname>Nogueira-Rodríguez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Dominguez-Carbajales</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>López-Fernández</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Iglesias</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Cubiella</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Fdez-Riverola</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Reboiro-Jato</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Glez-Pena</surname>
          </string-name>
          ,
          <article-title>Deep neural networks approaches for detecting and classifying colorectal polyps</article-title>
          ,
          <source>Neurocomputing</source>
          <volume>423</volume>
          (
          <year>2021</year>
          )
          <fpage>721</fpage>
          -
          <lpage>734</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>N.</given-names>
            <surname>Ghatwary</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Ahmed</surname>
          </string-name>
          , E. Grisan,
          <string-name>
            <given-names>H.</given-names>
            <surname>Jalab</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Bidaut</surname>
          </string-name>
          ,
          <string-name>
            <surname>X.</surname>
          </string-name>
          <article-title>Ye, In-vivo barrett's esophagus digital pathology stage classification through feature enhancement of confocal laser endomicroscopy</article-title>
          ,
          <source>Journal of Medical Imaging</source>
          <volume>6</volume>
          (
          <year>2019</year>
          )
          <fpage>014502</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>N.</given-names>
            <surname>Ghatwary</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Ahmed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Ye</surname>
          </string-name>
          ,
          <article-title>Automated detection of barrett's esophagus using endoscopic images: a survey, in: Annual conference on medical image understanding and analysis</article-title>
          , Springer,
          <year>2017</year>
          , pp.
          <fpage>897</fpage>
          -
          <lpage>908</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>Y. S.</given-names>
            <surname>He</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. R.</given-names>
            <surname>Su</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X. L.</given-names>
            <surname>Zuo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y. Q.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <article-title>Application of artificial intelligence in gastrointestinal endoscopy</article-title>
          ,
          <source>Journal of digestive diseases 20</source>
          (
          <year>2019</year>
          )
          <fpage>623</fpage>
          -
          <lpage>630</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>P.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T. M.</given-names>
            <surname>Berzin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. R. G.</given-names>
            <surname>Brown</surname>
          </string-name>
          , S. Bharadwaj,
          <string-name>
            <given-names>A.</given-names>
            <surname>Becq</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Xiao</surname>
          </string-name>
          , P. Liu,
          <string-name>
            <given-names>L.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Song</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , et al.,
          <article-title>Real-time automatic detection system increases colonoscopic polyp and adenoma detection rates: a prospective randomised controlled study</article-title>
          ,
          <source>Gut</source>
          <volume>68</volume>
          (
          <year>2019</year>
          )
          <fpage>1813</fpage>
          -
          <lpage>1819</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>K.</given-names>
            <surname>Namikawa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Hirasawa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Yoshio</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Fujisaki</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Ozawa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Ishihara</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Aoki</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Yamada</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Koike</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Suzuki</surname>
          </string-name>
          , et al.,
          <article-title>Utilizing artificial intelligence in endoscopy: a clinician's guide</article-title>
          ,
          <source>Expert Review of Gastroenterology &amp; Hepatology</source>
          <volume>14</volume>
          (
          <year>2020</year>
          )
          <fpage>689</fpage>
          -
          <lpage>706</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>P.</given-names>
            <surname>Brandao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Zisimopoulos</surname>
          </string-name>
          , E. Mazomenos, G. Ciuti,
          <string-name>
            <given-names>J.</given-names>
            <surname>Bernal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Visentini-Scarzanella</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Menciassi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Dario</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Koulaouzidis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Arezzo</surname>
          </string-name>
          , et al.,
          <article-title>Towards a computed-aided diagnosis system in colonoscopy: automatic polyp segmentation using convolution neural networks</article-title>
          ,
          <source>Journal of Medical Robotics Research</source>
          <volume>3</volume>
          (
          <year>2018</year>
          )
          <fpage>1840002</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>Y. H.</given-names>
            <surname>Choi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y. C.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Hong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Kim</surname>
          </string-name>
          , H.
          <string-name>
            <surname>-H. Won</surname>
          </string-name>
          , T. Kim,
          <article-title>Centernet-based detection model and u-net-based multi-class segmentation model for gastrointestinal diseases</article-title>
          ., in: EndoCV@ ISBI,
          <year>2020</year>
          , pp.
          <fpage>73</fpage>
          -
          <lpage>75</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>S.</given-names>
            <surname>Rezvy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Zebin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Braden</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Pang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Taylor</surname>
          </string-name>
          , X. Gao,
          <article-title>Transfer learning for endoscopy disease detection and segmentation with mask-rcnn benchmark architecture</article-title>
          ,
          <source>in: 2020 IEEE 17th International Symposium on Biomedical Imaging</source>
          ,
          <volume>17</volume>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>L. D.</given-names>
            <surname>Huynh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Boutry</surname>
          </string-name>
          ,
          <article-title>A u-net++ with pre-trained eficientnet backbone for segmentation of diseases and artifacts in endoscopy images and videos</article-title>
          , volume
          <volume>2595</volume>
          ,
          <string-name>
            <surname>CEUR-WS</surname>
          </string-name>
          ,
          <year>2020</year>
          , pp.
          <fpage>13</fpage>
          -
          <lpage>17</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>A.</given-names>
            <surname>Subramanian</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Srivatsan</surname>
          </string-name>
          ,
          <article-title>Exploring deep learning based approaches for endoscopic artefact detection and segmentation</article-title>
          , volume
          <volume>2595</volume>
          ,
          <string-name>
            <surname>CEUR-WS</surname>
          </string-name>
          ,
          <year>2020</year>
          , pp.
          <fpage>51</fpage>
          -
          <lpage>56</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>Y. B.</given-names>
            <surname>Guo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Zheng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B. J.</given-names>
            <surname>Matuszewski</surname>
          </string-name>
          ,
          <article-title>Deep encoder-decoder networks for artefacts segmentation in endoscopy images</article-title>
          , volume
          <volume>2595</volume>
          ,
          <string-name>
            <surname>CEUR-WS</surname>
          </string-name>
          ,
          <year>2020</year>
          , pp.
          <fpage>18</fpage>
          -
          <lpage>21</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>L. F.</given-names>
            <surname>Sánchez-Peralta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Bote-Curiel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Picón</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F. M.</given-names>
            <surname>Sánchez-Margallo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. B.</given-names>
            <surname>Pagador</surname>
          </string-name>
          ,
          <article-title>Deep learning to find colorectal polyps in colonoscopy: A systematic literature review</article-title>
          ,
          <source>Artificial Intelligence In Medicine</source>
          (
          <year>2020</year>
          )
          <fpage>101923</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>K.</given-names>
            <surname>Cho</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B. Van</given-names>
            <surname>Merriënboer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Bahdanau</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Bengio</surname>
          </string-name>
          ,
          <article-title>On the properties of neural machine translation: Encoder-decoder approaches</article-title>
          ,
          <source>arXiv preprint arXiv:1409.1259</source>
          (
          <year>2014</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>V.</given-names>
            <surname>Badrinarayanan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Kendall</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Cipolla</surname>
          </string-name>
          ,
          <article-title>Segnet: A deep convolutional encoder-decoder architecture for image segmentation</article-title>
          ,
          <source>IEEE transactions on pattern analysis and machine intelligence</source>
          <volume>39</volume>
          (
          <year>2017</year>
          )
          <fpage>2481</fpage>
          -
          <lpage>2495</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>S.</given-names>
            <surname>Ali</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Jha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Ghatwary</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Realdon</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Cannizzaro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. A.</given-names>
            <surname>Riegler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Halvorsen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Daul</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Rittscher</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O. E.</given-names>
            <surname>Salem</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Lamarque</surname>
          </string-name>
          , T. de Lange,
          <string-name>
            <given-names>J. E.</given-names>
            <surname>East</surname>
          </string-name>
          ,
          <article-title>Polypgen: A multi-center polyp detection and segmentation dataset for generalisability assessment</article-title>
          ,
          <source>arXiv</source>
          (
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>S.</given-names>
            <surname>Ali</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Braden</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Bailey</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Yang</surname>
          </string-name>
          , G. Cheng, P. Zhang,
          <string-name>
            <given-names>X.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Kayser</surname>
          </string-name>
          , R. D.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>