<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Robust 3D U-Net Segmentation of Macular Holes</article-title>
      </title-group>
      <contrib-group>
        <aff id="aff0">
          <label>0</label>
          <institution>Bioscience Institute, Newcastle University</institution>
          ,
          <addr-line>Newcastle Upon Tyne</addr-line>
          ,
          <country country="UK">UK</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Department of Computer Science, Durham University</institution>
          ,
          <addr-line>Durham</addr-line>
          ,
          <country country="UK">UK</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Gli .ai</institution>
          ,
          <addr-line>Durham</addr-line>
          ,
          <country country="UK">UK</country>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>School of Computing, Newcastle University</institution>
          ,
          <addr-line>Newcastle Upon Tyne</addr-line>
          ,
          <country country="UK">UK</country>
        </aff>
        <aff id="aff4">
          <label>4</label>
          <institution>Sunderland Eye In rmary</institution>
          ,
          <addr-line>Sunderland</addr-line>
          ,
          <country country="UK">UK</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Macular holes are a common eye condition which result in visual impairment. We look at the application of deep convolutional neural networks to the problem of macular hole segmentation. We use the 3D U-Net architecture as a basis and experiment with a number of design variants. Manually annotating and measuring macular holes is time consuming and error prone, taking dozens of minutes to annotate a single 3D scan. Previous automated approaches to macular hole segmentation take minutes to segment a single 3D scan. We found that, in less than one second, deep learning models generate signi cantly more accurate segmentations than previous automated approaches (Jaccard index boost of 0:08 0:09) and expert agreement (Jaccard index boost of 0:13 0:20). We also demonstrate that an approach of architectural simpli cation, by greatly simplifying the network capacity and depth, results in a model which is competitive with state-of-the-art models such as residual 3D U-Nets.</p>
      </abstract>
      <kwd-group>
        <kwd>Machine learning</kwd>
        <kwd>image processing and computer vision</kwd>
        <kwd>medicine</kwd>
        <kwd>segmentation</kwd>
        <kwd>neural nets</kwd>
        <kwd>retina</kwd>
        <kwd>macular holes</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>
        Idiopathic full thickness macular holes (iFTMH) are a common, and visually
disabling condition, being bilateral in 10% of a ected individuals. They occur
at a prevalence of approximately 1 in 200 of the over 60-year-old population
with an incidence of approximately 4000 per annum in the United Kingdom
(UK)[
        <xref ref-type="bibr" rid="ref1 ref13">1,13</xref>
        ]. If left untreated they result in visual acuity below the de nition of
blindness and typically greater than 1.0 logMAR (logarithm of the minimum
angle of resolution), where 0.1 logMAR is classed as normal.
      </p>
      <p>
        3D high-resolution images of the retina can be created using optical
coherence tomography (OCT) [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. It is now the standard tool for diagnosing macular
Copyright 2021 for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0)
holes [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. Compared to previous imaging methods, OCT can more easily assist a
clinician in di erentiating a full-thickness macular hole from mimicking
pathology, which is important in de ning appropriate treatment [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. An OCT scan of
a macular hole is a 3D volume. Clinicians, however, typically view OCT scans
as a series of 2D images, choose the central slice with maximum dimensions and
perform measurements which are predictors of anatomical and visual success
such as base diameter, macular hole inner opening and minimum linear
diameter [
        <xref ref-type="bibr" rid="ref12 ref15 ref19 ref2">12,2,15,19</xref>
        ]. This approach is limited as it assumes that the macular hole
base is circular, and would give incorrect results when it is elliptical [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ], which
is typically the case [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. With the advent of automated 3D approaches, it is
possible to begin to look at measurements in 3D and how they might be predictors
of anatomical and visual success.
      </p>
      <p>
        Neural networks are an interconnected group of arti cial neurons, which can
be recon gured to solve a problem based on data. Convolutional neural networks
(CNN) are a type of neural network inspired by how the brain processes visual
information [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]. CNNs have been very successful in computer vision problems,
such as automating the segmentation of medical images. For a CNN to learn to
segment images in a supervised manner, it needs to have access to images with
associated ground truth (GT) information which highlight the areas of the image
for the task at hand. This is often done manually which is time consuming and
requires expert knowledge.
      </p>
      <p>
        The U-Net CNN architecture [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ] is a highly utilized CNN architecture for
biomedical image segmentation for use on 2D images. It has had success in
segmentation to help diagnose other eye conditions such as macular edema, even
when dataset sizes are limited [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. We sought to examine the application of
variants of the U-Net architecture to the problem of macular hole segmentation. Our
proposed model is a smaller version of the model from the original 3D U-Net
paper [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. We also implemented and evaluated the proposed model with residual
blocks added, similar to those described by He et al. [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. In addition, we
implemented a much more complex residual model, DeepMind's OCT segmentation
model [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], and ran the same tests with it.
      </p>
      <p>
        Alternatives to U-Net have been created such as V-Net [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] which uses 3D
convolutions and a Dice score-based loss. We use binary cross-entropy as our loss
function, similar to the weighted cross entropy used in the original 3D U-Net
paper. Early experiments showed that binary cross-entropy outperformed a Dice
score-based loss for our problem. Additionally, a study that did a comparison
between multiple model architectures on another biomedical image
segmentation problem showed that V-Net-based models did not outperform U-Net-based
models [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. For these reasons, we chose 3D U-Net as the basis of our model rather
than V-Net.
      </p>
      <p>
        Our contribution can be summarized as developing an automated approach to
macular hole segmentation based on deep learning which yields signi cantly
improved results compared to prior methods. We present a comparison of the
abovementioned models against the current state-of-the-art automated approach [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ].
      </p>
      <p>The state-of-the-art method is a level set approach which does not use deep
learn160x188x49x1
80x94x24x64
80x94x24x128
= Conv3d, batchnorm3d + ReLU
= ConvTranspose3d
= Up convolution
= Max pooling
= Skip connection
40x47x12x128</p>
      <p>40x47x12x256
ing. We show that simple low-capacity 3D U-Nets are capable of outperforming
the state-of-the-art automated approach and that increasing the complexity of
the architecture does not improve performance. The PyTorch-based code for this
work has been released as an open-source project 6.</p>
    </sec>
    <sec id="sec-2">
      <title>Materials and Methods</title>
      <sec id="sec-2-1">
        <title>Materials</title>
        <p>All had undergone Spectral domain optical coherence tomography (SDOCT)
imaging using the Heidelberg Spectralis (Heidelberg, Germany) as part of routine
care, using the same imaging protocol. A high density central horizontal scanning
protocol with 29-30 micron line spacing was used in the central 15 by 5 degrees.
The individual OCT line scans were 768 496 pixels with the scaling varying
slightly between datasets but typically equating to 5.47 microns per pixel in the
x (horizontal) axis and 3.87 microns per pixel in the y (vertical) axis. With
2930 microns spacing between scans (z axis), there were 49 scans per dataset. All
scans used a 16 automatic real time setting enabling multisampling and noise
reduction over 16 images. All scans collected were from unique patients and were
stored using the uncompressed TIFF le format.</p>
        <p>All images were cropped to the same size and unnecessary information such
as the fundus image were removed. Annotations were created by a mixture of
clinicians and image experts using a 3D image annotation tool. Pixels on each
slice of the OCT scan which represented macular hole were highlighted. There
6 https://github.com/gli -ai/robust-3d-unet-macular-holes
were 85 (image, annotation) pairs in the training dataset, 56 after combining
annotations from multiple authors. There were 22 pairs in the validation dataset
and 9 in the unseen test set.</p>
        <p>Originally we had three annotations for each OCT image in the unseen test
set. However, due to inconsistencies between authors, we combined all ground
truths into a single ground truth per image. To do this, we used a voting system,
where if 32 of the authors had annotated a voxel, that voxel was annotated in the
resultant ground truth. All images and ground truths at full size had dimensions
321 376 49. We did not augment our dataset as we found that augmentations
did not improve the generalizability of our model. As we believe that our test and
validations sets are large enough to be representative of the real-world problem,
this was not deemed to be an issue.
2.2</p>
      </sec>
      <sec id="sec-2-2">
        <title>Methods</title>
        <p>Image segmentation involves the labelling of objects of interest in an image.
For a 3D image, this is done by assigning voxels with shared characteristics to
corresponding class labels. We wished to assign areas of the macular hole volume
in an OCT image to white voxels and all other regions to black voxels.</p>
        <p>We used binary cross-entropy as our loss function, which tells us how close
our predicted macular hole regions are to those in the ground truth:
LBCE =</p>
        <p>N
1 X pi log qi + (1
N i=1
pi) log(1
qi);
(1)
N being the batch size, pi being the ground truth and qi being the output of our
model. For images with multiple annotations in our training set, we trusted them
with equal integrity and the target probabilities were averaged. The validation
set had no samples with multiple annotations. As described in Section 2.1, for
the unseen test set, we used a voting process to decide on the nal target ground
truth.</p>
        <p>
          U-Net takes as input a 2D image and outputs a set of probabilities. Each
entry in the output is the probability of each part of the image being a part
of the segmented region. It is a U-shaped CNN architecture, consisting of a
contracting path and an expansive path. The contracting path consists of 2D
convolutions, ReLU activations and 2D max pooling at each level. The expansive
path's levels use skip connections to their contracting path equivalent, along with
2D convolutions, ReLU activations and 2D up convolutions. Skip connections
allow for high-resolution information to be captured by the model while the
contracting/expansive paths capture the abstract shape of the segmentation.
The 3D U-Net architecture [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ] is a version of U-Net designed for use with 3D
images which uses 3D convolutions, up convolutions and max pooling layers.
This allows for improved segmentation of 3D images as the context from multiple
slices are used to decide whether an individual voxel is an object or not.
        </p>
        <p>A number of models based on the 3D U-Net architecture were compared:</p>
        <sec id="sec-2-2-1">
          <title>M1: Small 3D U-Net (Proposal) [5,216,353 parameters],</title>
          <p>
            M2: Small residual 3D U-Net (Residual) [13,928,833 parameters],
M3: Residual 3D U-Net for 2D slices (DeepMind) [
            <xref ref-type="bibr" rid="ref4">4</xref>
            ] [470,333,089 parameters].
          </p>
          <p>A diagram of model M1 is shown in Fig. 1. Early on, versions of 2D U-Net
were implemented as described in the original paper, however, performance was
very poor for our dataset. The original 2D U-Net model has an input size of
572 572 and an output size of 388 388. The poor performance we noticed is
likely due to a lack of context from multiple slices. In addition, the input to the
original 2D U-Net model is of a higher resolution than our image slices, which
have a resolution of 321 376. Therefore, we needed to upscale our images which
resulted in distortion and wasted memory usage. Similarly, we also implemented
3D U-Net as described in the original paper, however, this also performed poorly.
Again, this is primarily due to the input and output sizes of the model being too
dissimilar to our dataset's. The original 3D U-Net model has an input size of
132 132 116. Our images only have 49 slices, which needed to be upscaled to
116 slices for this. This resulted in signi cant distortion of the input. The output
of the original 3D U-Net model is also of a very low resolution: 44 44 28
which would have been very coarse when upscaled to our image dimensions of
321 376 49. As our images were of resolution 321 376 49, we aimed to keep
the resolution of the input and output as close to this as possible. Di erent to
the original U-Net and 3D U-Net papers, it was decided to keep the input and
output dimensions equal to each other, to maximise the resolution of our output.
We tweaked convolution sizes, padding and strides until we achieved this goal,
while still tting in available GPU memory.</p>
          <p>
            Our experiments showed that using three levels for this model resulted in the
best performance, rather than the four levels that the original 3D U-Net paper
used. A scaled-down input image of 160 188 49 yielded the best results for
models M1 and M2. The output is of the same dimensions as the input. M2 is
similar to M1 except that residual blocks have been added to each level. M3 is
a very deep residual 3D U-Net architecture which takes nine slices of the OCT
image as input and outputs a 2D probability map as output, representing the
segmentation of a single slice of the OCT image. For M3, the slice which we
want to segment, along with 4 slices on either side is input to the model, which
is a 321 376 9 image. This is based on a model architecture developed by
DeepMind for segmenting OCT images [
            <xref ref-type="bibr" rid="ref4">4</xref>
            ]. For slices near the boundaries, we
use mirroring to handle slices that are outside of the image. It outputs a set of
321 376 probabilities, corresponding to one slice of the 3D OCT. M3, therefore,
requires 49 iterations to segment a whole 3D OCT image in our dataset. Model
M3 has the most parameters of the models tested, with M1 having the fewest
parameters.
          </p>
          <p>
            The Jaccard index was used as the primary metric for measuring the
performance of each method. This is one of the standard measures of the performance
of image segmentation methods, especially in medical image segmentation [
            <xref ref-type="bibr" rid="ref20">20</xref>
            ].
The Dice similarity coe cient (DSC) is another commonly used metric and is
closely related to the Jaccard index, with one being computable from the other.
          </p>
          <p>For completeness and ease of comparison with other results, we also provide the
DSC for our proposed model in Section 4.2.
3</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Implementation</title>
      <p>
        Our experiments were all conducted using the Python programming language
and the PyTorch [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ] deep learning framework on NVIDIA Turing GPUs with
24GB of memory. PyTorch is a state-of-the-art framework for building deep
learning models which is highly optimized for modern GPU hardware. We trained
each model for 500 epochs where each epoch ran over 10 3D images, which was
enough for all models to stop substantially improving. This means that the
models which output a 3D segmentation (M1 and M2) had 10 iterations per
epoch, and the slice-based model (M3) had 490 iterations per epoch. As source
code was not released for DeepMind's model, M3 was implemented as closely as
possible to the description provided in the original paper and slightly adapted
to t the binary classi cation problem.
      </p>
      <p>
        In order to evaluate models M1 and M2, we scaled up the output probability
map to its original size using trilinear interpolation and thresholded it at 0:5
to generate a binary segmentation. For model M3, we individually ran over
all 49 slices of an image and recombined the 49 2D probability maps into a
single 3D probability map. We then thresholded this combined map at 0:5 to
generate a 3D binary segmentation. The Adam optimization algorithm [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] was
used to optimize parameters of the models, with hyperparameters being found
by experimentation. The BCEWithLogitsLoss function in PyTorch was used for
loss calculation, which combines a sigmoid activation and binary cross entropy
loss into one function. A similar number of experiments were conducted for each
model. For model M1, a learning rate of 1e 4 and weight decay of 1e 6 was
used. For model M2, a learning rate of 1e 4 and weight decay of 1e 5 was
used. For model M3, a learning rate of 7:5e 5 was used and weight decay was
disabled. The 3D OCT images were normalized to the [0; 1] range prior to scaling
or slicing.
      </p>
      <p>Each model was trained and evaluated separately three times to assess the
consistency of our results. We then calculated the Jaccard index, comparing each
of the models' predictions with the ground truth. Due to the fact that we only
had a small number of images with multiple authors, we decided to keep the
training, validation and unseen test sets static for all tests rather than using
k-fold cross-validation. We reserved all images which had three annotations for
the unseen test set, in order for us to be able to compare our results with expert
agreement, which was a key goal of the research.
4</p>
    </sec>
    <sec id="sec-4">
      <title>Results</title>
      <p>In this section, we look at evaluating our models both qualitatively and
quantitatively. For qualitative results, we primarily present results in 2D for ease of
comparison with other methods. We also present a sample of segmented macular
OCT scan</p>
      <p>Ground truth</p>
      <p>M1 (Proposal)M2 (Residual)M3 (DeepMind)
hole volumes in 3D to demonstrate that our method captures the 3D shape of
the volume. For quantitative results, we present an image-by-image comparison
of each model's performance using the Jaccard index against the state-of-the-art
method. We then present a variety of other image segmentation metrics on the
proposed model for ease of comparison with other methods.
4.1</p>
      <sec id="sec-4-1">
        <title>Qualitative Results</title>
        <p>The qualitative results of running inference on the trained macular hole models
are generally quite close to the ground truth, as seen in Fig. 2. In general,
predictions from all of the models are closer to the ground truth than the
state-of-theart automated approach. We can see that the qualitative di erence between the
models tested is not hugely signi cant. This is surprising as M3 has signi cantly
more capacity than M1. This shows that adding more capacity to a model of a
particular architecture does not necessarily yield an improvement in qualitative
output.</p>
        <p>
          3D visualizations of the output of our proposed model can be seen in Table 1.
We can see how the 3D shape of the macular holes is preserved, and matches
gures from similar works [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ]. This type of view would allow the clinician to
view the macular hole from every angle, rather than the 2D views which are
currently widely used.
4.2
        </p>
      </sec>
      <sec id="sec-4-2">
        <title>Quantitative Results</title>
        <p>Fig. 3 shows how the average Jaccard index on the unseen test set improved
as M1 was trained and we can see that after 200 epochs it had surpassed the
performance of the state-of-the-art automated approach and expert agreement.</p>
        <p>3D segmentation</p>
        <p>
          All of the trained macular hole models perform very well compared to the
state-of-the-art automated approach [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ] as we see in Table 2. Despite model M1
having by far the fewest parameters, it achieves performance which is similar to
the highest-capacity model, and in some cases surpasses it. Further results in
Table 3 show that M1 performs consistently well under other standard
segmentation quality measures.
5
        </p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Discussion</title>
      <p>The results show that previous automated approaches to this problem cannot
compete with deep learning methods. All of the models tested performed
significantly better than the level set method.</p>
      <p>If we examine the model results in isolation, we can see that the results
can be divided into two categories: the high-capacity 3D U-Net model (M3)
1
0:8
x
ed0:6
n
i
d
r
a
c
ca0:4
J
0:2
0
0</p>
      <sec id="sec-5-1">
        <title>M1 (Proposal)</title>
        <p>Nasrulloh
Expert 1 vs Expert 2
Expert 1 vs Expert 3
Expert 2 vs Expert 3
100
200
300
400</p>
        <sec id="sec-5-1-1">
          <title>Epoch</title>
          <p>and the lower-capacity 3D U-Nets (M1 and M2). The low-capacity 3D U-Nets
achieve the best results on the unseen test set. The high-capacity model, which
has many times the number of parameters of the M1 model, does not have
better generalizability. This is even more surprising given that the high-capacity
model takes the full-resolution image as input and also outputs a full-resolution
segmentation. Given that the low-capacity 3D U-Nets use a downsized 3D image
as input and output, we would expect them to perform worse due to not having
the same amount of information available. The fact that this does not occur
implies that the chosen models do not need very high-resolution input and output
to make an accurate segmentation of macular holes in OCT images.</p>
          <p>
            It is a counterintuitive nding that we do not see an improvement in
performance for a model which takes a full-resolution image and which has a signi
cantly higher capacity. In a similar problem, a high-pro le study used this
highcapacity model for their segmentation [
            <xref ref-type="bibr" rid="ref4">4</xref>
            ]. Since that work did not present results
from di erent architectures as we have done, it is di cult to know whether our
results would be replicated there. Our work clearly shows that for some
biomedical segmentation problems, it is important to consider lower-capacity models in
addition to higher-capacity models.
          </p>
          <p>
            Our work concentrates on looking at OCT images from a particular type
of device, from a single manufacturer. For future work, other models of OCT
device should be tested and compared with our results. It has been shown that
models trained on one device can be relatively easily trained to work with other
devices [
            <xref ref-type="bibr" rid="ref4">4</xref>
            ]. Our data is from a particular population centre, namely North East
England. For future work, it would be interesting to see if our results are
replicated in other population centres, both nationally and internationally. As we
have made our code available as an open-source project, it is hoped that this
can be achieved.
          </p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>Conclusions</title>
      <p>
        All of the models tested exceeded the performance of the state-of-the-art
automated approach which is a level set method. It is clear that deep learning
methods allow for the generation of segmentations that are closer to what humans
provide. Despite M3 having 90 times the parameters of M1, M1 gives excellent
qualitative and quantitative results which are of a similar quality to M3. M1's
performance exceeded expert agreement by a Jaccard index of 0:13 0:20. As
M1 is the smallest model, it requires the least amount of resources to run. M1 is
also a quick model to run, requiring only one pass through the whole 3D image,
whereas M3 requires one pass per slice. Once trained, M1 is capable of
segmenting an OCT image in less than one second. In contrast, the state-of-the-art
automated method requires minutes to run [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ]. For these reasons, M1 is the best
candidate to form the basis of future studies in a clinical setting. These ndings
show that careful tuning and in some cases architectural simpli cation can, for
some simple task distributions, be as e ective as very deep residual designs.
      </p>
      <p>The code is provided as an open-source project in order for future researchers
to replicate our results and build upon this research. Training and testing on
di erent populations with di erent demographics will be crucial to determine
that our trained models do not exhibit any bias. The lack of large-scale open data
sets from di erent population centres for OCT imagery makes this a signi cant
challenge that needs to be overcome.
7</p>
    </sec>
    <sec id="sec-7">
      <title>Con icts of Interest</title>
      <p>In accordance with his ethical obligation as a researcher, Jonathan Frawley
reports that he received funding for his PhD from Gli .ai. Some of the work
described was developed as part of his work as an employee at Gli .ai. Gli .ai
also provided annotations created by non-clinicians. Data and annotations by
the clinician for this project were kindly provided by Maged Habib, Caspar
Geenen and David H. Steel of the Sunderland Eye In rmary, South Tyneside and
Sunderland NHS Foundation Trust, UK. All images were collected as part of
routine care and anonymised.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Ali</surname>
            ,
            <given-names>F.S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Stein</surname>
            ,
            <given-names>J.D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Blachley</surname>
            ,
            <given-names>T.S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ackley</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Stewart</surname>
            ,
            <given-names>J.M.</given-names>
          </string-name>
          :
          <article-title>Incidence of and risk factors for developing idiopathic macular hole among a diverse group of patients throughout the United States</article-title>
          .
          <source>JAMA Ophthalmology</source>
          <volume>135</volume>
          (
          <issue>4</issue>
          ),
          <volume>299</volume>
          {
          <fpage>305</fpage>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nasrulloh</surname>
            ,
            <given-names>A.V.</given-names>
          </string-name>
          , Wilson,
          <string-name>
            <given-names>I.</given-names>
            ,
            <surname>Geenen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            ,
            <surname>Habib</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            ,
            <surname>Obara</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            ,
            <surname>Steel</surname>
          </string-name>
          ,
          <string-name>
            <surname>D.H.</surname>
          </string-name>
          :
          <article-title>Macular hole morphology and measurement using an automated threedimensional image segmentation algorithm</article-title>
          .
          <source>BMJ Open Ophthalmology</source>
          <volume>5</volume>
          (
          <issue>1</issue>
          ),
          <year>e000404</year>
          (
          <year>2020</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Cicek</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Abdulkadir</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lienkamp</surname>
            ,
            <given-names>S.S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Brox</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ronneberger</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          :
          <article-title>3D U-Net: learning dense volumetric segmentation from sparse annotation</article-title>
          .
          <source>In: International Conference on Medical Image Computing and Computer-Assisted Intervention</source>
          . pp.
          <volume>424</volume>
          {
          <issue>432</issue>
          (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>De Fauw</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ledsam</surname>
            ,
            <given-names>J.R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Romera-Paredes</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nikolov</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tomasev</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Blackwell</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          , et al.:
          <article-title>Clinically applicable deep learning for diagnosis and referral in retinal disease</article-title>
          .
          <source>Nature Medicine</source>
          <volume>24</volume>
          (
          <issue>9</issue>
          ),
          <volume>1342</volume>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Frawley</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Willcocks</surname>
            ,
            <given-names>C.G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Habib</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Geenen</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Steel</surname>
            ,
            <given-names>D.H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Obara</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          :
          <article-title>Segmentation of macular edema datasets with small residual 3D U-Net architectures</article-title>
          .
          <source>In: 2020 IEEE 20th International Conference on Bioinformatics and Bioengineering (BIBE)</source>
          . pp.
          <volume>582</volume>
          {
          <issue>587</issue>
          (
          <year>2020</year>
          ). https://doi.org/10.1109/BIBE50027.
          <year>2020</year>
          .00100
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Ghavami</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hu</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gibson</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bonmati</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Emberton</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Moore</surname>
            ,
            <given-names>C.M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Barratt</surname>
            ,
            <given-names>D.C.</given-names>
          </string-name>
          :
          <article-title>Automatic segmentation of prostate MRI using convolutional neural networks: Investigating the impact of network architecture on the accuracy of volume measurement and MRI-ultrasound registration</article-title>
          .
          <source>Medical Image Analysis</source>
          <volume>58</volume>
          ,
          <issue>101558</issue>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Goldberg</surname>
            ,
            <given-names>R.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Waheed</surname>
            ,
            <given-names>N.K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Duker</surname>
            ,
            <given-names>J.S.:</given-names>
          </string-name>
          <article-title>Optical coherence tomography in the preoperative and postoperative management of macular hole and epiretinal membrane</article-title>
          .
          <source>British Journal of Ophthalmology</source>
          <volume>98</volume>
          (
          <issue>Suppl 2</issue>
          ),
          <year>ii20</year>
          {
          <fpage>ii23</fpage>
          (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>He</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhang</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ren</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sun</surname>
          </string-name>
          , J.:
          <article-title>Identity mappings in deep residual networks</article-title>
          .
          <source>In: European Conference On Computer Vision</source>
          . pp.
          <volume>630</volume>
          {
          <issue>645</issue>
          (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Hee</surname>
            ,
            <given-names>M.R.</given-names>
          </string-name>
          , Pulia to, C.A.,
          <string-name>
            <surname>Wong</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Duker</surname>
            ,
            <given-names>J.S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Reichel</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Schuman</surname>
            ,
            <given-names>J.S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Swanson</surname>
            ,
            <given-names>E.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fujimoto</surname>
            ,
            <given-names>J.G.</given-names>
          </string-name>
          :
          <article-title>Optical coherence tomography of macular holes</article-title>
          .
          <source>Ophthalmology</source>
          <volume>102</volume>
          (
          <issue>5</issue>
          ),
          <volume>748</volume>
          {
          <fpage>756</fpage>
          (
          <year>1995</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Kingma</surname>
            ,
            <given-names>D.P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ba</surname>
          </string-name>
          , J.:
          <article-title>Adam: A method for stochastic optimization</article-title>
          . In: Bengio,
          <string-name>
            <given-names>Y.</given-names>
            ,
            <surname>LeCun</surname>
          </string-name>
          , Y. (eds.) International Conference on Learning Representations (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Lindsay</surname>
            ,
            <given-names>G.W.</given-names>
          </string-name>
          :
          <article-title>Convolutional neural networks as a model of the visual system: Past, present, and future</article-title>
          .
          <source>Journal of Cognitive</source>
          Neuroscience p.
          <volume>1</volume>
          {
          <issue>15</issue>
          (Feb
          <year>2020</year>
          ). https://doi.org/10.1162/jocn a 01544
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Madi</surname>
            ,
            <given-names>H.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Masri</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Steel</surname>
            ,
            <given-names>D.H.</given-names>
          </string-name>
          :
          <article-title>Optimal management of idiopathic macular holes. Clinical Ophthalmology (Auckland</article-title>
          , NZ)
          <volume>10</volume>
          ,
          <issue>97</issue>
          (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>McCannel</surname>
            ,
            <given-names>C.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ensminger</surname>
            ,
            <given-names>J.L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Diehl</surname>
            ,
            <given-names>N.N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hodge</surname>
            ,
            <given-names>D.N.</given-names>
          </string-name>
          :
          <article-title>Population-based incidence of macular holes</article-title>
          .
          <source>Ophthalmology</source>
          <volume>116</volume>
          (
          <issue>7</issue>
          ),
          <volume>1366</volume>
          {
          <fpage>1369</fpage>
          (
          <year>2009</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Milletari</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Navab</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ahmadi</surname>
            ,
            <given-names>S.A.:</given-names>
          </string-name>
          <article-title>V-Net: Fully convolutional neural networks for volumetric medical image segmentation</article-title>
          .
          <source>In: 2016 Fourth International Conference on 3D Vision (3DV)</source>
          . pp.
          <volume>565</volume>
          {
          <issue>571</issue>
          (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Murphy</surname>
            ,
            <given-names>D.C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nasrulloh</surname>
            ,
            <given-names>A.V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lendrem</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Graziado</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Alberti</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>la Cour</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Obara</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Steel</surname>
            ,
            <given-names>D.H.</given-names>
          </string-name>
          :
          <article-title>Predicting postoperative vision for macular hole with automated image analysis</article-title>
          .
          <source>Ophthalmology Retina</source>
          (
          <year>2020</year>
          ). https://doi.org/10.1016/j.oret.
          <year>2020</year>
          .
          <volume>06</volume>
          .005, http://www.sciencedirect.com/ science/article/pii/S2468653020302311
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <surname>Nasrulloh</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Willcocks</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jackson</surname>
            ,
            <given-names>P.T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Geenen</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Habib</surname>
            ,
            <given-names>M.S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Steel</surname>
            ,
            <given-names>D.H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Obara</surname>
            ,
            <given-names>B.:</given-names>
          </string-name>
          <article-title>Multi-scale segmentation and surface tting for measuring 3D macular holes</article-title>
          .
          <source>IEEE Transactions on Medical Imaging</source>
          <volume>37</volume>
          (
          <issue>2</issue>
          ),
          <volume>580</volume>
          {
          <fpage>589</fpage>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          17.
          <string-name>
            <surname>Paszke</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gross</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Massa</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          , et al.:
          <article-title>PyTorch: an imperative style, highperformance deep learning library</article-title>
          .
          <source>In: Advances in Neural Information Processing Systems</source>
          <volume>32</volume>
          , pp.
          <volume>8024</volume>
          {
          <issue>8035</issue>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          18.
          <string-name>
            <surname>Ronneberger</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fischer</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Brox</surname>
          </string-name>
          , T.:
          <article-title>U-Net: convolutional networks for biomedical image segmentation</article-title>
          .
          <source>In: Medical Image Computing and Computer-Assisted Intervention</source>
          . vol.
          <volume>9351</volume>
          , pp.
          <volume>234</volume>
          {
          <issue>241</issue>
          (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          19.
          <string-name>
            <surname>Steel</surname>
            ,
            <given-names>D.H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Donachie</surname>
            ,
            <given-names>P.H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Aylward</surname>
            ,
            <given-names>G.W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Laidlaw</surname>
            ,
            <given-names>D.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Williamson</surname>
            ,
            <given-names>T.H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yorston</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>Factors a ecting anatomical and visual outcome after macular hole surgery: ndings from a large prospective UK cohort</article-title>
          . Eye pp.
          <volume>1</volume>
          {
          <issue>10</issue>
          (
          <year>2020</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          20.
          <string-name>
            <surname>Taha</surname>
            ,
            <given-names>A.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hanbury</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Metrics for evaluating 3D medical image segmentation: analysis, selection, and tool</article-title>
          .
          <source>BMC Medical Imaging</source>
          <volume>15</volume>
          (
          <issue>1</issue>
          ),
          <volume>1</volume>
          {
          <fpage>28</fpage>
          (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>