<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Land Cover Semantic Segmentation Using ResUNet</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Loukas Kouvaras</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Eleni Charou</string-name>
          <email>exarou@iit.demokritos.gr</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Vasilis Pollatos</string-name>
          <email>vaspoll97@gmail.com</email>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Harokopio University</institution>
          ,
          <addr-line>Athens</addr-line>
          ,
          <country country="GR">Greece</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>NCSR Demokrtios</institution>
          ,
          <addr-line>Athens</addr-line>
          ,
          <country country="GR">Greece</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>NTUA</institution>
          ,
          <addr-line>Athens</addr-line>
          ,
          <country country="GR">Greece</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>In this paper we present our work on developing an automated system for land cover classification. This system takes a multiband satellite image of an area as input and outputs the land cover map of the area at the same resolution as the input. For this purpose convolutional machine learning models were trained in the task of predicting the land cover semantic segmentation of satellite images. This is a case of supervised learning. The land cover label data were taken from the CORINE Land Cover inventory and the satellite images were taken from the Copernicus hub. As for the model, U-Net architecture variations were applied. Our area of interest are the Ionian islands (Greece). We created a dataset from scratch covering this particular area. In addition, transfer learning from the BigEarthNet dataset [1] was performed. In [1] simple classification of satellite images into the classes of CLC is performed but not segmentation as we do. However, their models have been trained into a dataset much bigger than ours, so we applied transfer learning using their pretrained models as the first part of out network, utilizing the ability these networks have developed to extract useful features from the satellite images (we transferred a pretrained ResNet50 into a U-Res-Net). Apart from transfer learning other techniques were applied in order to overcome the limitations set by the small size of our area of interest. We used data augmentation (cutting images into overlapping patches, applying random transformations such as rotations and flips) and cross validation. The results are tested on the 3 CLC class hierarchy levels and a comparative study is made on the results of diferent approaches.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>LULC, U-NET, deep learning, transfer learning,Ionio</p>
    </sec>
    <sec id="sec-2">
      <title>INTRODUCTION</title>
      <p>Modern AI technologies, such as deep learning, can be utilized in
various fields of natural science to automate and underpin
procedures traditionally carried out by humans. Remote sensing
nowadays provides a great amount of data of high quality which are
updated on a daily basis. Another important thing is that these
data are easily produced and are open to the public in contrast to
other sources, such as aerial photography that are of higher quality
but are more expensively and less massively produced. For some
problems (in our case land cover recognition) the resolution of the
open remote sensing data (10m for sentinel-2) is adequate. The big
data of remote sensing can be fed into machine learning models
to develop automated systems that analyse this data and carry out
useful tasks. Labeled data are the most useful ones, as they can be
utilised for the purposes of supervised learning that solves a great
range of problems.</p>
      <p>CLC provides a huge labeled dataset. It contains maps for the
most part of Europe for the last three decades. Our goal is to train
models to predict the labels of the CLC dataset. Most research done
in this field is about assigning one or more land cover labels into a
whole satellite image patch (which can take an area of several square
kilometres). Our approach to the problem is more general, trying
to construct a semantic segmentation of the satellite image into the
full range of the land cover classes provided by the Corine Land
Cover inventor, at the maximal resolution provided by sentinel-2
satellite images, which is 10m. The classes of CLC are hierarchical.
We are testing the ability of the models to predict the classes on
each one of the hierarchical levels. As expected, we see that the
superclasses on the higher levels are discriminated with greater
accuracy than the subclasses on the lower levels.</p>
      <p>Corine Land Cover has a wide variety of applications,
underpinning various Community policies in the domains of environment,
but also agriculture, transport, spatial planning. Developing a
system that automates the production of CLC maps to some extent
is important because CLC needs to be updated every few years.
Creating these maps is a burdensome and time-consuming job for
the human and even so the accuracy of the produced maps isn’t
perfect. An automatic land cover classification system could help
develop such maps in the future, track down sudden or short term
changes that happen to the land cover (for example due to natural
disasters or due to fast track rural and urban development). It could
also be applied to areas that are not included in the CLC.</p>
      <p>State of the art deep learning models were used and the training
and testing were done in the area of Ionio. This is a case of work on
a relatively small area with special geological and natural features.
It is also an area of varying morphology and landscapes and small
scale land cover characteristics that can hardly be detected on the
resolution provided by sent-2 images. Similar approaches can be
used for training and testing in other areas covered by the sentinel-2
satellites. As a first step we trained a simple U-Net from scratch in
the area of interest. Recently, a similar research was done in the TU
Berlin, developing the BigEarthNet. They perform simple
classification of satellite images into the classes of CLC but not segmentation
as we do. However, their models have been trained into a dataset
much bigger than ours, so we applied transfer learning using their
pretrained models as the first part of out network, utilizing the
ability these networks have developed to extract useful features
from the satellite images (we transferred a pretrained ResNet50 into
a U-Res-Net). Apart from transfer learning other techniques were
applied in order to overcome the limitations set by the small size of
our area of interest. We used data augmentation (cutting images
into overlapping patches, applying random transformations such
as rotations and flips) and cross validation.
is distributed over 6 Ionian islands (Corfu, Paxi, Lefkada, Kalamos,
Kefalonia, Zante) and the coast of Parga.
2</p>
    </sec>
    <sec id="sec-3">
      <title>RELATED WORK</title>
      <p>
        Land Cover Recognition gathers a lot of interest in the research
community. In our work we apply transfer learning from the
models trained in BigEarthNet [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. The BigEarthNet dataset contains
590,326 non-overlapping image patches of size 1200m ×1200m
distributed over 10 european countries (Austria, Belgium, Finland,
Ireland, Kosovo, Lithuania, Luxembourg, Portugal, Serbia,
Switzerland). Each image patch is annotated by multiple land-cover classes
(i.e., multi-labels) that are provided from the CORINE Land Cover
database of the year 2018 (CLC 2018). They train models that take
each patch as input and predict the classes appearing in this patch.
They solve a simpler problem than ours, because the resolution of
the output of their models is 1200m, while the resolution of our
predicted maps is 10m. However, their models have been trained on
a dataset much bigger than ours and have learned to extract useful
features from the images (encoding) that are later on decoded to
solve their task. We are using the pretrained encoder of a res-net-50
trained on BigEarthNet as the encoder part of a unet-like
architecture to solve our semantic segmentation problem. This approach
has also been adopted by [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. UNet architecture was introduced
in [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ]. ResNetUnet, the architecture we are using, is commonly
used for such problems. In [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] a sophisticated ResNetUnet that
performs multitasking achieved state of the art results for the ISPRS
2DPotsdam dataset. One of the subproblems solved in this
multitasking is finding the class boundaries, which is also proposed in
[
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]. However, as far as our problem is concerned, these methods
are applied on high resolution images of urban areas and may be
of little use for our problem. In order to conquer the limitations set
by our small dataset, data augmentation is applied as in [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ], [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ],
[
        <xref ref-type="bibr" rid="ref22">22</xref>
        ]. In our work we used Sentinel-2 bands with 10m resolution
and bands with 20m resolution. Others have used multisource data
including optical data and Sentinel-1 radar measurements [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] ,[
        <xref ref-type="bibr" rid="ref15">15</xref>
        ],
[
        <xref ref-type="bibr" rid="ref16">16</xref>
        ]. Multi-temporal data viewing the same area on diferent
timestamps is another approach taken in [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ],[
        <xref ref-type="bibr" rid="ref17">17</xref>
        ],[
        <xref ref-type="bibr" rid="ref18">18</xref>
        ],[
        <xref ref-type="bibr" rid="ref19">19</xref>
        ]. In order to
deal with missing labels active learning [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ],[
        <xref ref-type="bibr" rid="ref20">20</xref>
        ], self-learning [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ]
and weakly supervised learning [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] is performed.
3
3.1
      </p>
    </sec>
    <sec id="sec-4">
      <title>METHODOLOGY</title>
    </sec>
    <sec id="sec-5">
      <title>Dataset</title>
      <p>Our dataset was created by multispectral satellite images of the
Ionian Islands downloaded from Copernicus for the period of 2018
and part of the CLC 2018 that covers the Ionian Islands. CLC
vector files were georeferenced together with the Copernicus images,
turned into raster with 10m resolution and altogether were clipped
in the same bounds creating tifs for each one of the islands. These
tifs were cut into patches of size 1,28x1,28 km (128x128 pixels) with
some high degree of overlap. Xdata consists of these patches having
the satellite image bands as features for each pixel and Ydata
consists of the corresponding CLC patches. Our networks are trained
to solve the task of predicting the CLC label for each pixel of the
input patch, given the band measurements for each pixel of the
input patch. So we are trying to find a function f such that Ydata =
f (Xdata). This is a case of supervised learning. Our area of interest</p>
      <sec id="sec-5-1">
        <title>Kalamos</title>
      </sec>
      <sec id="sec-5-2">
        <title>Parga</title>
      </sec>
      <sec id="sec-5-3">
        <title>South Corfu</title>
      </sec>
      <sec id="sec-5-4">
        <title>North Corfu</title>
      </sec>
      <sec id="sec-5-5">
        <title>Kefalonia</title>
      </sec>
      <sec id="sec-5-6">
        <title>Lefkada</title>
      </sec>
      <sec id="sec-5-7">
        <title>North Zante</title>
      </sec>
      <sec id="sec-5-8">
        <title>Paxi</title>
        <p>For each area we have the sentinel-2 10m resolution bands (R,G,B,
infrared), the sentinel-2 20m resolution bands (b05, b06, b07, b8A,
b11 and b12) and the corine land cover class label for each pixel. In
our problem the satellite image bands are the inputs to our network
and the clc classes the expected output.</p>
        <p>Corine Land Cover classes are hierarchical into three levels. Our
approach is training the models on the full range of the corine land
cover classes and then testing them on each level separately.</p>
        <p>The area of interest has to be splitted into training and test sets.
Due to the small size of our dataset we chose not to use a validation
set for the fine tuning of hyperparameters such as the number of
epochs. The training process was stopped when the loss function
started to converge and not when it was minimal for the validation
set. We are performing cross validation so the area of interest has
to be divided into a number of subsets of approximately same size .
The area of interest was partitioned into the following 6 subsets: 1.
north Corfu, 2. south Corfu, 3. west Kefalonia, 4. east Kefalonia, 5.
Lefkada, 6. Paxi+North Zante+Kalamos+Parga The splitting into
training and validation sets is done 6 times, so that each time a
diferent subset is the validation set and the remaining 5 are the
training set.</p>
        <p>Each area is cut into overlapping patches. The overlaps are a
form of data augmentation. Patch size is 1.28 km x 1.28 km and
the hop between adjacent patches is 0.64 km in each direction
(longitude and latitude). Two memory optimisations were applied.
Firstly, patches are stored by defining only their limits in the original
satellite image and the cutting is only performed on dataloading.
Secondly, patches containing only sea are discarded ( e.g. the blue
square in the right image below). This is a good practice because
it turns out that the models are able to learn to recognise the sea
almost perfectly even without those patches. It also reduces class
imbalancement, as sea patches are the most frequent ones in our
area.</p>
        <p>For the transfer learning experiments data needed to be
standardised using the same mean and std values as the base model.
On dataloading random flips and rotations were applied for the
purposes of data augmentation.</p>
      </sec>
      <sec id="sec-5-9">
        <title>Overlapping patches 1.28 km x 1.28 km patch</title>
        <p>Two diferent approaches were followed. The first approach was
to train a baseline UNet from scratch into the area of interest. The
second approach was to perform transfer learning. The transfer
learning UNet model has a ResNet-50 architecture on the encoder
part and the weights of the encoder are initialised to the values
of the weights of a ResNet-50 trained on the BigEarthNet . The
ifgure below shows the exact architecture of the transfer learning
model. There are approximately 66.000.000 trainable parameters on
this model. A more complex version of this model that applied no
compression on the outputs of the encoder that were passed to the
decoder through shortcuts had 91.000.000 trainable parameters and
improved fitting on the training set but didn’t seem to generalise
better than the model presented below.</p>
        <p>The baseline UNet model that was trained from scratch solved
an easier problem, as the output and the ground truth land cover
images had a resolution of 100m.
3.3</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>Training</title>
      <p>We are trying to solve a semantic segmentation problem and a
composite dice and a binary cross entropy loss with logits criterion
is used.The two loss criteria are summed, each one with a weight
factor of 0.5. We experimented with positive weights pc in the bce:
 (, ) =  = {1, , . . . ,  , }⊤,</p>
      <p>, = −, , · log  (, ) + (1 − , ) · log(1 −  (, ))
where c is the class number.</p>
      <p />
      <p>Setting  = (      ) , for diferent
values of a in (0, 1] for class balancing deteriorated our results.
Adam optimiser is used to achieve fitting in the training data. initial
 = 5 · 10−4 and it gradually decreases with the use of a
scheduler.The complexity of our model requires the use of regularization
techniques. We applied dropout, with rate 0-0.2 for the outer layers
and 0.3-0.4 for the inner hidden layers. For the first epochs of the
training, the weights of the base transfer learning model remain
frozen. We unfreeze them when the learning process starts to
converge, dropping at the same time the learning rate. As we can see
below, unfreezing the base model on epoch 80 causes some
instability. However, after some epochs the loss returns to the low values
it had before the unfreezing. The pretrained encoder seems to work
properly without further training, but the unfreezing brings some
slight improvements so we perform it. Training was executed on
google colab.
Several versions of the problem are being examined. Firstly, training
from scratch was done on the area of interest. A baseline model
shown in figure 6 was used. The produced maps had a resolution
100m. The visual results and the metrics for the validation set are
presented below:</p>
      <sec id="sec-6-1">
        <title>Kefalonia (target)</title>
      </sec>
      <sec id="sec-6-2">
        <title>Kefalonia (prediction)</title>
        <p>class
1.1 Urban fabric
1.2 Industrial,
commercial and transport units
1.3 Mine, dump and
construction sites
1.4 Artificial,
nonagricultural vegetated
areas
2.1 Arable land
2.2 Permanent crops
2.3 Pastures
2.4 Heterogeneous
agricultural areas
3.1 Forest
3.2 Shrub and/or
herbaceous vegetation
associations
3.3 Open spaces with
little or no vegetation
4.1 Inland wetlands
5.2 Marine waters
accuracy = 0.85329
 1 = 0.4124
 1 = 0.85329
 1ℎ = 0.8522
0.0003858
0.02698
0.186
0.081
0.6165</p>
        <p>Classification Report
2.Agricultural 3.Forest
areas and
seminatural
areas
1922887 2332193
0.7729 0.8075
0.747 0.8537
class
1.1.2 Discontinuous
urban fabric
1.3.1 Mineral extraction
sites
1.4.2 Sport and leisure
facilities
2.1.1 Non-irrigated
arable land
2.2.3 Olive groves
2.3.1 Pastures
2.4.2 Complex
cultivation patterns
2.4.3 Land principally
occupied by agriculture,
with significant areas of
natural vegetation
3.1.2 Coniferous forest
3.1.3 Mixed forest
3.2.1 Natural grassland
3.2.3 Sclerophyllous
vegetation
3.2.4 Transitional
woodland/shrub
3.3.2 Bare rock
3.3.3 Sparsely vegetated
areas
5.2.3 Sea and ocean
accuracy = 0.88019
 1 = 0.559
 1 = 0.88019
 1ℎ = 0.89214
Fold 3:
class</p>
        <p>In the experiments presented above our method was to keep
a continuous area as a validation set, for example a whole island.
Now we present a diferent approach where the validation patches
are randomly distributed over the area of interest. This is also a
realistic problem, where the experts sparsely assign land cover
labels on the area of interest and the remaining unlabeled areas are
predicted by a model trained on the neighbouring labeled ones. To
make sure that the training and the validation set have no common
elements we skipped data augmentation via overlaps, but the flips
and rotations are still used. We split the area of interest into train
and validation with a ratio of 70, 30 respectively.</p>
        <p>The metrics for the validation are presented below.
 1 = 0.28225
 1 = 0.59871
 1ℎ = 0.62438</p>
        <p>Finally we are going to present some examples that show the
performance of our model. All the predictions presented are on
validation data.The number on the top of each image on the left
indicates the fold number (6-fold cross validation).</p>
        <p>In some of the above examples we see the dificulty of our
problem, deriving from the low resolution of the input images and the
ambiguity of the corine labels. In some cases the model made the
right predictions, even though it is a dificult task even for the
human observing the rgb input image.
• Our models provide a basis for the creation of land cover
maps based on the CLC nomenclature. The visual results
show the ability of our models to find the boundaries
between classes and the accuracy on the higher levels of the
class hierarchy is pretty good. The accuracy on common
subclasses is also good. However, the performance on predicting
uncommon classes and discriminating subclasses of the same
superclass on the lower levels of the CLC class hierarchy
isn’t adequate and human supervision may be needed for
this task.
• The CLC dataset contains imperfections. These limit the
accuracy of our models. However, in some cases the model
can outperform the accuracy of the dataset in cases where
the dataset has a lower quality than it’s average.
• Usually the land cover is mixed or can not be described
accurately by the existing CLC classes. This leads to discord
between the labeled data and the predictions, even for kinds
of land cover that have been seen on the training set. We
also observe that sometimes there are multiple class labels
that could describe the land cover and despite the seeming
disagreement between the model output and the labels they
are close to each other. This indicates the need for a more
sophisticated loss criterion and performance metrics that give
diferent penalties to diferent types of confusion between
classes, taking into account the hierarchical structure of the
classes and the similarities and overlaps between classes.
• Increasing the resolution of the output from 100m to 10m
can give better results but bigger models are required (more
parameters).
• The main contribution of transfer learning was speeding up
the training processes and possibly improving the results.
The encoder part of the network didn’t have to be trained,
at least for the first epochs of the training, resulting in
decreased epoch duration.
• Using a bigger dataset could boost the performance of our
models in the area of interest, especially in the task of
predicting uncommon classes.</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>G.</given-names>
            <surname>Sumbul</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Charfuelan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Demir</surname>
          </string-name>
          , and
          <string-name>
            <given-names>V.</given-names>
            <surname>Markl Bigearthnet</surname>
          </string-name>
          :
          <article-title>A large-scale benchmark archive for remote sensing image understanding</article-title>
          .
          <source>IEEE International Conference on Geoscience and Remote Sensing Symposium</source>
          , pp.
          <fpage>5901</fpage>
          -
          <lpage>5904</lpage>
          ,
          <year>Jul 2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>Copernicus</given-names>
            <surname>Programme. CORINE Land</surname>
          </string-name>
          <article-title>Cover</article-title>
          . Last Accessed:
          <fpage>2020</fpage>
          -10-10.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>Priit</given-names>
            <surname>Ulmas</surname>
          </string-name>
          , Innar Liiv,
          <article-title>Segmentation of Satellite Imagery using U-Net Models for Land Cover Classification</article-title>
          . arXiv:
          <year>2003</year>
          .02899
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <surname>Charou</surname>
          </string-name>
          , Eleni, George Felekis, Danai Bournou Stavroulopoulou, Maria Koutsoukou, Antigoni Panagiotopoulou, Yorghos Voutos, Emmanuel Bratsolis, Phivos Mylonas, and
          <string-name>
            <surname>Laurence</surname>
          </string-name>
          Likforman-Sulem.
          <article-title>Deep Learning for Agricultural Land Detection in Insular Areas</article-title>
          .
          <source>In 2019 10th International Conference on Information, Intelligence, Systems and Applications (IISA)</source>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>4</lpage>
          . IEEE,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <surname>Diakogiannis</surname>
            ,
            <given-names>Foivos</given-names>
          </string-name>
          &amp; Waldner, Francois &amp; Caccetta,
          <string-name>
            <surname>Peter</surname>
          </string-name>
          &amp; Wu,
          <string-name>
            <surname>Chen.</surname>
          </string-name>
          (
          <year>2019</year>
          ).
          <article-title>ResUNet-a: a deep learning framework for semantic segmentation of remotely sensed data</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>A.</given-names>
            <surname>Nivaggioli</surname>
          </string-name>
          and
          <string-name>
            <given-names>H.</given-names>
            <surname>Randrianarivo</surname>
          </string-name>
          ,
          <article-title>Weakly supervised semantic segmentation of satellite images</article-title>
          .
          <source>CoRR</source>
          , vol. abs/
          <year>1904</year>
          .03983,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>S.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. M.</given-names>
            <surname>Xie</surname>
          </string-name>
          , G. Azzari, and
          <string-name>
            <given-names>D. B.</given-names>
            <surname>Lobell</surname>
          </string-name>
          ,
          <article-title>Weakly supervised deep learning for segmentation of remote sensing imagery</article-title>
          .
          <source>Remote Sensing</source>
          , vol.
          <volume>12</volume>
          , no.
          <issue>2</issue>
          , p.
          <fpage>207</fpage>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>J.</given-names>
            <surname>Ahn</surname>
          </string-name>
          and
          <string-name>
            <surname>S.</surname>
          </string-name>
          <article-title>Kwak Learning pixel-level semantic afinity with image-level supervision for weakly supervised semantic segmentation</article-title>
          .
          <source>CoRR</source>
          , vol. abs/
          <year>1803</year>
          .10464,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <surname>Mohammadimanesh</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Salehi</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mahdianpari</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gill</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Molinier</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          (
          <year>2019</year>
          )
          <article-title>A new fully convolutional neural network for semantic segmentation of polarimetric SAR imagery in complex land cover ecosystem</article-title>
          .
          <source>ISPRS Journal of Photogrammetry and Remote Sensing</source>
          ,
          <volume>151</volume>
          ,
          <fpage>223</fpage>
          -
          <lpage>236</lpage>
          . https://doi.org/10.1016/j.isprsjprs.
          <year>2019</year>
          .
          <volume>03</volume>
          .015
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>Dimitrios</surname>
            <given-names>Marmanis</given-names>
          </string-name>
          , Konrad Schindler, Jan Dirk Wegner, Silvano Galliani, Mihai Datcu,
          <article-title>Uwe Stilla Classification With an Edge: Improving Semantic Image Segmentation with Boundary Detection</article-title>
          . arXiv:
          <volume>1612</volume>
          .
          <fpage>01337</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>M.</given-names>
            <surname>Buchhorn</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Smets</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Bertels</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Lesiv</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.-E.</given-names>
            <surname>Tsendbazar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Herold</surname>
          </string-name>
          , and S.
          <source>Fritz “Copernicus Global Land Service: Land Cover 100m: epoch</source>
          <year>2015</year>
          : Globe,” Oct.
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>G. J.</given-names>
            <surname>Scott</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. R.</given-names>
            <surname>England</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W. A.</given-names>
            <surname>Starms</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. A.</given-names>
            <surname>Marcum</surname>
          </string-name>
          and
          <string-name>
            <given-names>C. H.</given-names>
            <surname>Davis Training</surname>
          </string-name>
          <article-title>Deep Convolutional Neural Networks for Land-Cover Classification of HighResolution Imagery IEEE Geoscience and Remote Sensing Letters</article-title>
          , vol.
          <volume>14</volume>
          , no.
          <issue>4</issue>
          , pp.
          <fpage>549</fpage>
          -
          <lpage>553</lpage>
          ,
          <year>April 2017</year>
          , doi: 10.1109/LGRS.
          <year>2017</year>
          .
          <volume>2657778</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Benbahria</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. F.</given-names>
            <surname>Smiej</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Sebari</surname>
          </string-name>
          , and
          <string-name>
            <given-names>H.</given-names>
            <surname>Hajji</surname>
          </string-name>
          <article-title>Land cover intelligent mapping using transfer learning and semantic segmentation</article-title>
          .
          <source>2019 7th Mediterranean Congress of Telecommunications (CMT)</source>
          , pp.
          <fpage>1</fpage>
          -
          <issue>5</issue>
          ,
          <year>10 2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>N.</given-names>
            <surname>Kussul</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Lavreniuk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Skakun</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <given-names>A.</given-names>
            <surname>Shelestov</surname>
          </string-name>
          <article-title>Deep learning classification of land cover and crop types using remote sensing data</article-title>
          .
          <source>IEEE Geoscience and Remote Sensing Letters</source>
          , vol.
          <source>PP</source>
          , pp.
          <fpage>1</fpage>
          -
          <issue>5</issue>
          ,
          <fpage>03</fpage>
          <lpage>2017</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>Y. J. E.</given-names>
            <surname>Gbodjo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Ienco</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Leroux</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Interdonato</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Gaetano</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Ndao</surname>
          </string-name>
          , and S. Dupuy “
          <article-title>Object-based multi-temporal and multi-source land cover mapping leveraging hierarchical class relationships</article-title>
          ,”
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <surname>Gbodjo</surname>
            ,
            <given-names>Yawogan</given-names>
          </string-name>
          &amp; Leroux, Louise &amp; Gaetano, Rafaele &amp; Ndao,
          <string-name>
            <surname>Babacar.</surname>
          </string-name>
          (
          <year>2019</year>
          ).
          <article-title>RNN-based Multi-Source Land Cover mapping: An application to West African landscape</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <surname>Ziheng</surname>
            <given-names>Sun</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>Liping</given-names>
            <surname>Di</surname>
          </string-name>
          &amp; Hui
          <string-name>
            <surname>Fang</surname>
          </string-name>
          (
          <year>2019</year>
          )
          <article-title>Using long short-term memory recurrent neural network in land cover classification on Landsat and Cropland data layer time series</article-title>
          .
          <source>International Journal of Remote Sensing</source>
          ,
          <volume>40</volume>
          :
          <fpage>2</fpage>
          ,
          <fpage>593</fpage>
          -
          <lpage>614</lpage>
          , DOI: 10.1080/01431161.
          <year>2018</year>
          .1516313
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <surname>Kim</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Park</surname>
          </string-name>
          , N.-W.;
          <string-name>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <surname>K.-D.</surname>
          </string-name>
          Self
          <article-title>-Learning Based Land-Cover Classification Using Sequential Class Patterns from Past Land-Cover Maps</article-title>
          .
          <source>Remote Sens</source>
          .
          <year>2017</year>
          ,
          <volume>9</volume>
          ,
          <fpage>921</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>B.</given-names>
            <surname>Demir</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Bovolo</surname>
          </string-name>
          and
          <string-name>
            <given-names>L.</given-names>
            <surname>Bruzzone</surname>
          </string-name>
          ,
          <article-title>Updating Land-Cover Maps by Classification of Image Time Series: A Novel Change-Detection-Driven Transfer Learning Approach</article-title>
          .
          <source>in IEEE Transactions on Geoscience and Remote Sensing</source>
          , vol.
          <volume>51</volume>
          , no.
          <issue>1</issue>
          , pp.
          <fpage>300</fpage>
          -
          <lpage>312</lpage>
          , Jan.
          <year>2013</year>
          , doi: 10.1109/TGRS.
          <year>2012</year>
          .
          <volume>2195727</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <surname>Caleb</surname>
            <given-names>Robinson</given-names>
          </string-name>
          , Anthony Ortiz, Kolya Malkin, Blake Elias, Andi Peng, Dan Morris, Bistra Dilkina,
          <article-title>Nebojsa Jojic Human-Machine Collaboration for Fast Land Cover Mapping</article-title>
          . arXiv:
          <year>1906</year>
          .04176
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <surname>Ronneberger</surname>
          </string-name>
          , Olaf; Fischer, Philipp; Brox, Thomas (
          <year>2015</year>
          ).
          <article-title>"U-Net: Convolutional Networks for Biomedical Image Segmentation"</article-title>
          ..
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <surname>Stivaktakis</surname>
            ,
            <given-names>Radamanthys</given-names>
          </string-name>
          &amp; Tsagkatakis, Grigorios &amp; Tsakalides,
          <string-name>
            <surname>Panagiotis.</surname>
          </string-name>
          (
          <year>2019</year>
          ).
          <article-title>Deep Learning for Multilabel Land Cover Scene Categorization Using Data Augmentation</article-title>
          .
          <source>IEEE Geoscience and Remote Sensing Letters. PP. 1-5</source>
          .
          <fpage>10</fpage>
          .1109/LGRS.
          <year>2019</year>
          .
          <volume>2893306</volume>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>