<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Generic Semantic Segmentation of Historical Maps</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Rémi Petitpierre</string-name>
          <email>remi.petitpierre@epfl.ch</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Frédéric Kaplan</string-name>
          <email>frederic.kaplan@epfl.ch</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Isabella di Lenardo</string-name>
          <email>isabella.dilenardo@epfl.ch</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Digital Humanities Laboratory, EPFL</institution>
          ,
          <addr-line>Lausanne</addr-line>
          ,
          <country country="CH">Switzerland</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Institute for Area and Global Studies, EPFL</institution>
          ,
          <addr-line>Lausanne</addr-line>
          ,
          <country country="CH">Switzerland</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2021</year>
      </pub-date>
      <issue>4</issue>
      <fpage>228</fpage>
      <lpage>248</lpage>
      <abstract>
        <p>Research in automatic map processing is largely focused on homogeneous corpora or even individual maps, leading to inflexible models. Based on two new corpora, the first one centered on maps of Paris and the second one gathering maps of cities from all over the world, we present a method for computing the figurative diversity of cartographic collections. In a second step, we discuss the actual opportunities for CNN-based semantic segmentation of historical city maps. Through several experiments, we analyze the impact of figurative and cultural diversity on the segmentation performance. Finally, we highlight the potential for large-scale and generic algorithms. Training data and code of the described algorithms are made open-source and published with this article.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;historical map processing</kwd>
        <kwd>neural networks</kwd>
        <kwd>semantic segmentation</kwd>
        <kwd>computer vision</kwd>
        <kwd>topology</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        representational flexibility [
        <xref ref-type="bibr" rid="ref24">24</xref>
        ]) of graphical objects produced in diverse technical and cultural
contexts remain largely unsolved. In the following sections, we experimentally map the design
space for such a generic processing pipeline and present first working prototypes tested on two
large map corpora.
      </p>
      <sec id="sec-1-1">
        <title>1.1. Previous Works</title>
        <p>
          Classical segmentation algorithms are based on the specific knowledge of a map collection.
Despite color appearing relatively late in map printing processes [
          <xref ref-type="bibr" rid="ref46">47</xref>
          ], this graphical component
is frequently used. Some specialists even consider color-based pre-processing an “essential” task
[
          <xref ref-type="bibr" rid="ref36">36</xref>
          ]. The simplest paradigm in this regard is color thresholding [
          <xref ref-type="bibr" rid="ref12 ref15">12, 15</xref>
          ]. Other studies add a
morphological approach, using region growing algorithms [
          <xref ref-type="bibr" rid="ref30 ref31 ref9">9, 31, 30</xref>
          ], or rely on human feedback
[
          <xref ref-type="bibr" rid="ref11">11</xref>
          ]. In deep learning methods, region growing is also used to flood fill and add a semantic layer
to extracted polygons using watershed [
          <xref ref-type="bibr" rid="ref40">41</xref>
          ]. Another salient graphical component is texture.
Hatched areas are particularly targeted [
          <xref ref-type="bibr" rid="ref38 ref4 ref55">56, 4, 39</xref>
          ]. Other methods focus on texture energy
[
          <xref ref-type="bibr" rid="ref34">34</xref>
          ]. These approaches in general have met some success on textured maps [
          <xref ref-type="bibr" rid="ref36 ref50 ref54">36, 51, 55</xref>
          ]. Unlike
color, however, textures have many degrees of freedom, such as size and rotation. Their use
for segmentation therefore requires fine parametrization.
        </p>
        <p>
          Beyond pure graphical marks, some researches focus on detecting morphological features,
such as lines [
          <xref ref-type="bibr" rid="ref34">34</xref>
          ], or closed polygons [
          <xref ref-type="bibr" rid="ref35">35</xref>
          ]. These approaches are generally confronted with the
problem of incomplete lines, due to the degradation of the document or graphical choices (e.g.
dashed lines) and therefore require the development of reconstruction algorithms [
          <xref ref-type="bibr" rid="ref2 ref25 ref33 ref48">33, 2, 25,
49</xref>
          ]. Moreover, the extraction of the geometries is sensitive to information overlay, which is
extremely frequent in cartography. Many specific methods were therefore developed to detect
and eventually remove every disruption, such as the map grid [
          <xref ref-type="bibr" rid="ref26">26</xref>
          ], the background texture
[
          <xref ref-type="bibr" rid="ref34">34</xref>
          ], the text [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ], or the symbols [
          <xref ref-type="bibr" rid="ref56">57</xref>
          ]. The detection of these interfering elements generally
requires precise knowledge of their nature and their visual characteristics: size, shape, texture,
color, etc [
          <xref ref-type="bibr" rid="ref19">19</xref>
          ]. To conclude, the extraction of the morphology is still unable to add a semantic
layer to the map content. For instance, a rectangle might well be a building, but it can equally
well be a courtyard or a basin.
        </p>
        <p>
          More recently, CNNs opened up new perspectives for the resolution of semantic segmentation
problems [
          <xref ref-type="bibr" rid="ref32">32</xref>
          ]. In 2019, a first successful transfer was presented with the segmentation of parcels
from ”Napoleonic” cadastral maps of Venice [
          <xref ref-type="bibr" rid="ref40">41</xref>
          ], using a UNet architecture with a ResNet
encoder. In 2020, further results were obtained on the extraction of the railway network from
some USGS maps, using a modified PSPNet as encoder [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ]. Heitzler and Hurni also presented
a research on the extraction of building footprints from the Swiss Siegfried national maps
(1872-1949) [
          <xref ref-type="bibr" rid="ref21">21</xref>
          ]. Other works were focusing on extracting specific elements, such as places of
archaeological interest [
          <xref ref-type="bibr" rid="ref17">17</xref>
          ], or road types [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ]. This research question being topical, the ICDAR
2021 conference dedicated a competition to the segmentation of building blocks on a corpus
of maps from the Bibliothèque historique de la Ville de Paris [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ]. One solution in particular
stood out by proposing to use a DenseNet [
          <xref ref-type="bibr" rid="ref23">23</xref>
          ] architecture.
        </p>
        <p>
          On the other hand, [
          <xref ref-type="bibr" rid="ref53">54</xref>
          ] proposed a method to operationalize cartographic figuration based
on color-histogram based moments. The vector projection of these descriptors allows one to
efficiently visualize the figurative conventions of a map or a corpus of maps. Subsequently,
a texture extraction method has also been proposed [
          <xref ref-type="bibr" rid="ref52">53</xref>
          ]. Other eforts to operationalize
cartographic figuration have also been carried out concerning graphic map load, i.e. the visual
density of cartographic content [
          <xref ref-type="bibr" rid="ref3 ref51">3, 52</xref>
          ]. These researches are part of the field of information
theory.
        </p>
      </sec>
      <sec id="sec-1-2">
        <title>1.2. Research approach</title>
        <p>In this research, we seek to better understand the impact of figuration and figurative diversity
on the learning of neural networks, in the context of the semantic segmentation of historical
city maps. Thus, our research questions are the following: 1) How to measure the figurative
diversity in a map corpus? 2) How robust are CNNs when facing high figurative diversity?</p>
        <p>To answer these research questions, we first present a method to operationalize the
cartographic figuration and measure the figurative diversity of a map corpus. We demonstrate the
significant variability of our data, in comparison with other map corpora commonly found in
the literature. Then, we propose an efective processing pipeline involving pre-segmentation of
the map frame and semantic segmentation of the map itself. We highlight the potentials and
the limitations of neural networks for solving generic semantic segmentation problems. Finally,
we conduct a set of experiments that allow to investigate the learning mechanisms deployed by
neural networks and to explore the design space of map semantic segmentation. Specifically,
we seek to qualify the impact of learning transfer on each class. We then aim to evaluate the
perspectives for an educated corpus constitution, through examination of cross-cultural
performance and confidence prediction. Finally, we challenge the importance of graphical cues,
in comparison to non-graphical concepts.</p>
      </sec>
      <sec id="sec-1-3">
        <title>1.3. Dataset</title>
        <p>To create an experimental context that challenges the genericity of the semantic segmentation,
we contrast two corpora that present a diferent kind of variability. The first corpus gathers
330 maps of Paris from the collections of the Bibliothèque nationale de France (BnF) and the
Bibliothèque historique de la Ville de Paris. Most of the maps were published between 1800 and
1950 in the Paris region, and their scale is generally comprised between 1:25’000 and 1:2’000.
The second corpus gathers 256 maps of cities from all over the world, including numerous
reference maps. They come from 32 diferent collections, the main ones being: the BnF,
the Library of Congress, the Harvard Library, the David Rumsey Collection, the University
of Bordeaux, the British Library, the Boston Public Library and the Institut Cartogràfic i
Geològic de Catalunya. In total, the corpus represents 182 diferent cities in 90 countries.
The distribution across the various regions is balanced, except for Oceania (8 maps only).
The regions with the most maps are Eastern Europe &amp; Central Asia (34), Western Europe
(34), and East Asia (30). Conversely, the regions accounting for the fewest maps are Oceania,
Subsaharian Africa (15), and Middle East (17). The urban form of each map in the World
corpus was manually classified into three categories: regular (95 maps), irregular (68), or mixed
(93). Most maps were published between 1720 and 1950. However, due to historical reasons,
most non-Western publications occurred after 1800.</p>
        <p>Both corpora present specific difficulties. For the Paris corpus (Fig. 1), one of the most
complex issue to apprehend is the information cluttering or overlay. In particular, the
superimposition of information concerning mobility, such as road or underground network (Fig. 1
A.2-3), but also on the water system, the catacombs, or more simply on administrative
divisions (Fig. 1 B.3-4). For Paris, this intensive use of the city map as a tool for planning
urban works was likely caused by the lack of a proper cadastre before the late 19th century.
Low-contrast may also make the map images difficult to read (Fig. 1 B.3, C.4). For the World</p>
        <p>
          E
Figure 2: Map samples from the World corpus.
the frame, i.e. to non-geographic content of the document; and the map content, i.e. all the
geographic content that do not belong to the road network. In the extended ontology, the
map content was subdivided into 3 additional classes: the blocks, the water, and the non-built,
which in fact includes all non-aquatic unbuilt land, except the road network, i.e. wasteland,
meadows, crops, forests, but also parks, or inner courtyards. The Tab. 1 summarizes the
distribution of classes in the diferent corpora and sets. The training patches are open-source
and freely available online [
          <xref ref-type="bibr" rid="ref44">45</xref>
          ].
        </p>
        <p>
          As in [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ], we consider that the problem of the segmentation of the map frame is a diferent
problem, mainly for scale reasons. This step is therefore carried out beforehand, and the
presegmented background class is indicated as +1 in the next pages. The pre-segmentation simply
takes the form of a mask applied on the input images
        </p>
      </sec>
    </sec>
    <sec id="sec-2">
      <title>2. Methods</title>
      <sec id="sec-2-1">
        <title>2.1. Operationalization of the figuration</title>
        <p>
          To make cartographic figuration measurable, we extracted 3 sets of descriptors related to color,
texture, and orientation. The color and texture features are based on previous works by Uhl
et al. [
          <xref ref-type="bibr" rid="ref52 ref53">54, 53</xref>
          ]. First, mean and standard deviation are computed on the distribution of each
color channel. Then, 256-bins histograms are extracted for each of the latter and the skewness
and kurtosis in the distribution are computed. The mean is used to transcribe the hue and
the color value, the standard deviation indicates the contrasts. The skewness is a descriptor
of the asymmetry of the color distribution, while the kurtosis is a flattening coefficient of the
curve and therefore also allows to describe the contrasts. Color standard deviation, kurtosis
and skewness are summed for the 3 channels. Second, local binary patterns (LBP, [
          <xref ref-type="bibr" rid="ref39">40</xref>
          ]) are
computed, using a radius of 2, on the Otsu-binarized images [
          <xref ref-type="bibr" rid="ref42">43</xref>
          ]. A 4-bins histogram is
extracted on the LBP values. LBP are invariant to color value (i.e. brightness) and rotation.
They can help diferentiate between edges, corners or flat surfaces at a local scale and thus
characterize the texture at a larger scale. Third, a 24-bins histogram of oriented gradient
(HOG, [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ]) is computed. Each bin corresponds to a certain orientation angle of the image
local gradients. Therefore, they are ultimately grouped into 5 categories, summarizing the
orientation of the gradients: vertical (±π), horizontal (±π/2, ±3π/2), diagonal (±π/4, ±3π/4),
regular oblique (±π/6, ±2π/6, ±4π/6, ±5π/6), and irregular oblique (all other orientations).
At this point, there are 14 features in total: 6 (3+3) color descriptors, 4 LBP descriptors, and
5 HOG descriptors. Together, these descriptors can characterize the cartographic figuration,
as can be seen in Fig. 3.
        </p>
        <p>
          These visual features are computed, using 50x50 sub-patches, for each image of both corpora,
as well as for the USGS [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ], the Napoleonic cadaster [
          <xref ref-type="bibr" rid="ref40">41</xref>
          ], and the ICDAR21 dataset [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ], for
comparison purposes. On the one hand, the inter-corpora inter-class correlation is computed
between the Paris and World datasets, as well as the intra-corpus inter-class correlation within
each corpus.
        </p>
        <p>
          On the other hand, a 32-bins histogram is computed on the distribution of each feature.
As the distributions can be multimodal, due to repetitive homogeneous figuration, the modes
are extracted by smoothing the histogram with a Savitzky-Golay filter [
          <xref ref-type="bibr" rid="ref49">50</xref>
          ] of width 3 and
polynomial degree 1. The local minima are identified on a window of width 3 and the histogram
is split between each mode. To investigate the first research question, we want to determine
how much these features vary in a map corpus. To this extend, a κ-coefficient is defined as the
proportionally weighted sum of the kurtosis on each mode of the distribution. In other terms,
the κ-coefficient is a measure of the acuteness of the feature distribution in a map corpus, and
can thus characterize the homogeneity of the figuration. As the value of the κ-coefficient can
vary according to the size of the corpus, the bigger sample sets were randomly downsampled,
without replacement, to the size of the smallest sample set (here the Napoleonic cadaster set).
The κ-coefficient was then computed 5000 times for various downsampling schemes and for
each feature, the median κ-coefficient was retained as an estimator of the real κ-coefficient.
The bias of this recalibration is below ±3.6% for the World, and below ±2.0% for Paris, with
a confidence of 95%. The code of the described algorithms is made open-source and published
with this article [
          <xref ref-type="bibr" rid="ref43">44</xref>
          ].
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Map segmentation</title>
        <p>
          For the semantic segmentation, we are using a CNN with UNet architecture [
          <xref ref-type="bibr" rid="ref47">48</xref>
          ] and ResNet
[
          <xref ref-type="bibr" rid="ref20">20</xref>
          ] as encoder, implemented in a Pytorch version of the open-source tool dhSegment [
          <xref ref-type="bibr" rid="ref27 ref41">42, 27</xref>
          ].
The batch size and learning rate parameters are optimized. The method and the results of
the tests are detailed in [
          <xref ref-type="bibr" rid="ref45">46</xref>
          ]. The selected parameters are a batch size of 1, as in the original
article by Long and Shelhammer [
          <xref ref-type="bibr" rid="ref32">32</xref>
          ], and a learning rate of 5 ∗ 10−5. The decoder weights are
initialized using Glorot and Bengio uniform method [
          <xref ref-type="bibr" rid="ref18">18</xref>
          ]. The training data are augmented
by side flip and upside-down flip, and by rotation ( rϵ [0, π]). The loss used is cross entropy.
The optimization relies on stochastic gradient descent (SGD). Unless otherwise specified, the
encoder weights are first pretrained on ImageNet [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ].
        </p>
        <p>The second pretraining is done in a crossed way, Paris being pretrained on the World, and
the World being pretrained on Paris, as described in the following subsection. The CNN is
then trained successively during 150 epochs on the Paris (2+1 and 4+1) datasets and on the
World (2+1 and 4+1) datasets. ResNet101 is used as encoder with LeakyReLU as activation.</p>
        <p>Three metrics are used to quantify performance: intersection over union (IoU), precision,
and recall. In a second step, the confusion matrices between the diferent classes are computed.
They are normalized regarding the proportion of pixels belonging to the class, according to
the ground-truth.</p>
      </sec>
      <sec id="sec-2-3">
        <title>2.3. Semantic segmentation experiments</title>
        <sec id="sec-2-3-1">
          <title>2.3.1. Impact of learning transfer on each class</title>
          <p>In a first experiment, the CNN with ResNet50 as encoder was trained during 100 epochs on the
Paris 5-classes-dataset, respectively on the World 5-classes-dataset. After the first training,
the network was re-trained on the Paris set, using the weights trained in the previous step on
the World corpus as initialization. Reciprocally, the World set was re-trained by initializing
this time the weights on the Paris corpus.
2.3.2. Analysis of cross-cultural performance and dataset design
To estimate the bias of the validation set, and to investigate cross-cultural performance, an
8fold cross validation was performed on the World 4+1 dataset. Each time, the CNN was trained
during 100 epochs as described in the subsection on the performance of semantic segmentation.</p>
        </sec>
        <sec id="sec-2-3-2">
          <title>2.3.3. Perspectives on confidence prediction</title>
          <p>In this experiment, we attempted to create an estimator of confidence at the patch scale. First,
a 10-fold cross validation was performed on the Paris 3-classes-dataset in order to estimate the
error on each patch of the training set. A simple ResNet50 encoder was used and trained
during 60 epochs each time. The output predictions from the kth-fold are compared with the
ground truth, and an accuracy map, in which the pixels take the value 1 if the prediction is
correct and 0 if the prediction is wrong, is created.</p>
          <p>A second network, identical to the first one, is then trained on 300 pairs of images and
accuracy maps, and validated on another 30 pairs. Instead of segmenting, the aim of this
network is to predict the accuracy map corresponding to the input image. The output of the
CNN prediction is classified using a global threshold, which is set to meet the mean accuracy of
the training set. The confidence index is defined as the patch accuracy, and the reference as the
accuracy previously measured by k-fold cross validation. In order to evaluate the confidence
prediction performance, an 8-fold cross validation was performed on the World 4+1 dataset.
Each time, the CNN was trained during 100 epochs with the same parameters as in 2.2.</p>
        </sec>
        <sec id="sec-2-3-3">
          <title>2.3.4. Importance of graphical cues for learning</title>
          <p>GRAY</p>
          <p>BINARY</p>
          <p>TEXTURELESS BIN.</p>
          <p>
            This experiment aims to determine what role color, texture, and morphology take in the
CNN performance, in contrast with non-graphical concepts. For this purpose, the images of
the training and validation sets are subjected to 4 diferent treatments (Fig. 4): reference,
gray, binary, and textureless binary. For the reference, the images are not modified in any
way. The second treatment is the gray treatment. In the latter, the RGB color channels of the
image are transformed into a grayscale. For the third treatment, the images are transformed
into a grayscale, then binarized [
            <xref ref-type="bibr" rid="ref42">43</xref>
            ]. For the fourth treatment, the images are transformed
into a grayscale, binarized, and texture is extracted with LBP [
            <xref ref-type="bibr" rid="ref39">40</xref>
            ] (r = 3). Finally, a second
Otsu thresholding is applied. The CNN is trained 5 times for 60 epochs separately on each of
the 4 datasets.
          </p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Results</title>
      <sec id="sec-3-1">
        <title>3.1. Operationalization of the figuration</title>
        <p>The median overall κ was computed on each dataset (Tab. 2). The result is very close for
the two studied corpora, while the Napoleonic cadaster and the ICDAR dataset are already
further away. The USGS map is in a diferent order of magnitude.</p>
        <p>Fig. 5 and Tab. 3 are the aggregated representations of the correlations of the figurative
features between and within both corpora. The frame class seems to be represented very
similarly in both corpora (ρ = 0.96, Fig. 5), while the non-built class is the most distant
(ρ = 0.81, Fig. 5). In the Paris corpus, the blocks class seems to stand out clearly from the
other classes (ρ¯Blocks = 0.775, Tab. 3, and Fig. 5), while in the World corpus, it is rather the
road network that stands out (ρ¯RoadNetwork = 0.888, Tab. 3, and Fig. 5). In general, however,
all classes are more distinct in the Paris corpus (ρ¯Mean = 0.842 for Paris, ρ¯Mean = 0.917 for
the World, Tab. 3).
k
oadR treowN
e
m
a
r
F
s
i
r
a
P
n
o
i
t
a
l
e
r
r
o
c
s
s
a
l
c
r
e
t
n
I
s
i
r
a
P
d
l
r
o
W
n
o
i
t
a
l
e
r
r
o
c
s
s
a
l
c
s
u
p
r
o
c
r
e
t
n
I
d
l
r
o
W
n
o
i
t
a
l
e
r
r
o
c
s
s
a
l
c
r
e
t
n
I
Frame Blocks</p>
        <p>Water Non-built</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Map segmentation</title>
        <p>The results of the third experiment are summarized in Tab. 4 and Fig. 6. Some example
prediction outputs can be observed in Fig. 7. The clearest finding is the consistent drop in
performance, when increasing the number of classes from 2+1 to 4+1. The mean IoU (mIoU)
on the 4+1 classes problems is not sufficient for reliable map segmentation. However, the mIoU
is good on both World and Paris 2+1 corpora. The drop is more noticeable for precision than
recall. The second clear diference occurs between the Paris and the World corpora, the first
performing better. Again, the disparity is mostly due to a low recall. It is worth noticing that
the top-50% of the World corpus performs very similarly to Paris average sample.</p>
        <p>For the Parisian corpus, most confusion occurs as non-built is predicted as blocks, or to a
lesser extent, as blocks are predicted as non-built, or as non-built is predicted as blocks. As it
is also noticeable in Tab. 4, non-built is by far the class performing worse in the 4+1-classes
problem. However, the water is the class sufering from the lowest precision score. In the
2+1classes problem, the blocks are sometimes classified as road network, which heavily impacts
the precision of the road network class.
Metric
IoU
Precision
Recall
n
o
i
t
c
i
d
e
r
P
t
li
u
b
n
o
N
r
e
t
a
W
k
oadR treowN
AVERAGE Paris 2+1</p>
        <p>AVERAGE World 2+1</p>
        <p>TOP50% World 2+1</p>
        <p>For the World corpus, water on the contrary benefits from a relatively high precision but a
very low recall. It is heavily confused with non-built and blocks. Non-built is still sometimes
predicted as blocks, but for this dataset, the contrary is more frequent. Both blocks and road
network classes sufer from a low precision.</p>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. Semantic segmentation experiments</title>
        <sec id="sec-3-3-1">
          <title>3.3.1. Impact of learning transfer on each class</title>
          <p>As one can see in Fig. 8, the performance in class segmentation shows significant disparities
between both datasets. If, in general, the Parisian corpus achieves much higher performances,
the World corpus seems to be better at recognizing the non-built lands. The transfer learning
is quite successful for the World corpus, when pretrained on Paris. The water (pvalue =
0.0063 &lt; 0.05) in particular seems to be better recognized. The transfer learning from the
World corpus to Paris is a bit less successful. However, the water class also demonstrates a
significant improvement (0.0076). Overall, the improvement is trending for the World corpus,
when pretrained on Paris (0.062).</p>
          <p>Frame</p>
          <p>Road network
World wrt. Paris</p>
          <p>Paris pretrained on World
wrt. Paris</p>
          <p>World pretrained on Paris
wrt. World
3.3.2. Analysis of cross-cultural performances and dataset design
In total, the mIoU over the 8 experiments is 0.6112, which is noticeably better than the 0.5595
score obtained in the previous section for the same 4+1 World set. That means that the
average performance is slightly better than the performance observed on the validation set.
The mIoU can also be computed on each patch separately, which corresponds to an average
of 0.5424. Fig. 9, shows a few examples of the high disparity between the top-50% and the
bottom-50% of the World corpus, which was already noticed in the results of the previous
experiment.</p>
          <p>Maps that have been published by a Western country score 0.5608, while other maps score
0.4911. The region of the city represented is also impactful, with Subsaharian (0.6322) and
North African (0.6280) cities scoring best, followed closely by Eastern Europe and Central Asia
(0.5931), South America (0.5913), Western Europe (0.5891), and North America (0.5711). At
the end of the line are the South Asian cities (0.3985). In the middle, one would find the Middle
East (0.5048), East Asia (0.4896), Oceania (0.4685), and Central America (0.4635). The urban
form also has a clear impact, as cities with a more regular (0.5453) or mixed (0.5535) urban
form score better than cities with an irregular (0.5088) urban form. This performance drop
is especially noticeable on the blocks class for regular (0.3548), mixed (0.3027), and irregular
(0.2179) urban form.</p>
        </sec>
        <sec id="sec-3-3-2">
          <title>3.3.3. Perspectives on confidence prediction</title>
          <p>The average mIoU over the 10-folds is 0.6993. For this third setting, the correlation between
the obtained confidence index and the reference is 0.571 ( pvalue = 1.2 ∗ 10−3) on the validation
set, which represents an intermediate to high dependency.
SUCCESSFUL</p>
          <p>SUCCESSFUL</p>
          <p>UNSUCCESSFUL
E
G
A
M
I</p>
        </sec>
        <sec id="sec-3-3-3">
          <title>3.3.4. Importance of graphical cues for learning</title>
          <p>The removal of color had noticeably almost no impact on performance. The median loss is
only 1.3% with regard to the reference mIoU. The binarization of the values resulted in a
drop of 7.2%. Finally, the disappearance of colors and textures led to a 10.4% decrease in
performance. This experiment thus shows that even when most graphical cues are removed,
most of the performance is conserved, and therefore that neural networks may also heavily rely
on more abstract reasonings for image segmentation.</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Discussion</title>
      <p>As measured through the operationalization of the figuration, both corpora, Paris and the
World, present a much greater figurative diversity than the other datasets used in the literature
(Tab. 2). This is good news and validates the interest of the studied datasets. The USGS is
massively less diverse than the other datasets, which is logical since it is a digital-born map.
The representation of the diferent elements is therefore perfectly codified and reproduced. The
ICDAR21 corpus also shows relatively little diversity. It is composed of plates published in
diferent years, but is still based on a single printed collection, and thus on a unified grammar.
The Napoleonic cadaster for its part is famous for the high level of formalization of cartographic
grammar. However, its manual execution explains a certain residual diversity. Finally, the two
studied corpora show a similar level of figurative diversity, although Paris stems from a single
cultural pool. This is consistent with the samples taken from this corpus, which show a great
variety of grammars, a high density of information, and above all a high level of technicality,
which also allows for an important figurative diversity.</p>
      <p>Regarding the performances of semantic segmentation, while the Paris corpus shows excellent
results which already indicate a potential for production (mIoU 0.8905, Tab. 4), the results of
the World corpus seem globally still perfectible (mIoU 0.8055). However, the best half
(top50%) of the World corpus presents performances very similar to the Paris corpus and may thus
also present an important automation potential (see also Fig. 9). The question therefore lies in
identifying this outperforming half in the large map collections, to open automatic vectorization
perspectives. As we demonstrated in the experiment on confidence prediction, the estimation
of a confidence index is possible and could probably solve this problem by identifying promising
maps beforehand. However, we consider that further research is still needed to maximize the
reliability of such an index and that other paradigms should be explored.</p>
      <p>The identification of the most promising maps can also be based on the results of
crosscultural validation. Indeed, we observed that maps with regular urban forms, like most colonial
cities, or mixed forms, like most European capitals, are better segmented on average. This
is consistent with the results on the importance of non-graphical cues for learning, which
emphasize the crucial importance of non-visual features, such as morphology, topology, and
semantic hierarchy, in CNN performance. In general, we notice that areas urbanized more
recently, such as Africa, perform well. This is consistent with the findings on urban form. The
poorer performance of South and East Asian cities can be explained on the one hand by atypical
visual elements, from a Western point of view, by the irregularity of Indian and Islamic urban
forms, as well as, in some cases, by the poor state of preservation of the documents. The little
performance of the Oceanian maps could be explained by the relatively low representation
of the region (around 3%) in the sample. The cross-cultural validation also highlights the
greater ease of segmenting maps published in the West, although, as mentioned earlier,
nonWestern maps are on average more recent. These cultural biases tend to argue for a diferential
treatment of non-Western maps (e.g. Fig. 2 A.3-4, C.1-2).</p>
      <p>The detailed analysis of the results of semantic segmentation brings many additional and
interesting insights. First, we notice the particular case of the water class. This class was the
least represented in the training corpora (Tab. 1), in particular for Paris, where it represents
less than 3% of the surface. The proportion lies close to the road network for the World
(12%). However, that class difers little from the other classes, figuratively (r = 0.94, Tab.
3). This explains its poor final performance ( IoUP aris = 0.47, IoUW orld = 0.33, Tab. 4). The
performance is higher for Paris. However, water seems to be mostly recognized by elimination of
other classes, which is expressed in poor precision (0.49), compared to recall (0.88). Moreover,
the water class is vulnerable to overfitting on the blue color, as demonstrated by the particular
case of the blueprint (Fig. 9), while the results on the importance of graphical cues for learning
show on the contrary that most of the performance relies on more abstract features. These
elements point to an imbalance of the water class in the original sample. This imbalance is not
specific to our corpus, since it is the result of geographical constraints. However, our research
provides a first solution to this limitation. Indeed, the experiment on the impact of learning
transfer on each class demonstrated that the water class could benefit from a significant transfer
(Fig. 8, +18% for Paris, +36% for the World). The constitution of large and diversified corpora
intended for pretraining therefore seems to be justified by this example.</p>
      <p>The second noteworthy result is that the non-built class scores much better for the World
(0.6528), compared to Paris (0.4235), even though the World generally scores lower. For the
World, however, the figuration of this class is not really outstanding (in average rpearson =
0.924). In reality, this class benefits (and sufers) from a catch-all efect. Indeed, a third (0.33)
of the water surfaces and nearly a quarter (0.23) of the blocks surfaces are wrongly classified as
non-built, while the non-built class itself does not confuse these classes (0.03, respectively 0.12).
This catch-all efect results in a relatively high recall, compared to Paris, while the impact on
precision remains little, as this class represents a large area (0.359) of the dataset. As it can
be seen in Fig. 6, this catch-all non-built class is the main reason for the underperformance
of the World dataset. Solutions to this problem might include separating this class into two
smaller and figuratively more specific classes.</p>
      <p>For Paris, this same non-built class obtains lower results, though quite close to those obtained
for water, for example. This class is less of a catch-all, as the balance between confusing and
being confused is shifted. In particular, the non-built is abundantly confused with blocks
(0.33), even though the figuration between those two is the least correlated (0.7). Conversely,
blocks are also sometimes confused with the non-built (0.17). This confusion is by far the most
important reason for this poor performance. Several explanations are possible. Notably, in
Paris, this class mainly describes parks, as well as courtyards. Compared to the World, the
topology of the non-built class is less outstanding. Some parks adopt forms that are quite
similar to building blocks, especially in Paris. Moreover, the spatial juxtaposition of inner
courtyards and buildings promotes confusion, especially in this direction, since an unrecognized
inner courtyard will be classified as a non-built confused with blocks. The Parisian corpus
also sufers from poor learning transfer from the World, which may be caused by a relatively
large figuration distance between Paris and the World for this class ( rpearson = 0.81). This
seems obvious, considering that the elements represented in one and the other corpus are also
semantically quite diferent, as explained.</p>
      <p>The arguments discussed above do not pretend to be exhaustive because neural networks
have many as yet unknown springs. However, they do shed some light on a few points. We
consider that the framework of the maps is ideal for this discussion on the ability of neural
networks to combine figurative and topological cues and to open up avenues of understanding
on their performances. Indeed, maps have a ”textbook” figuration, with basic textures, such
as hatching, some colors, and partly geometric morphology, squares, rectangles, trapezoids.
They are therefore easier to characterize from this point of view. These conditions are met
in very few fields of application of deep learning. In addition, for semantic segmentation, the
ability to visualize the results and understand the errors is particularly useful for interpretation.
This field therefore brings together all the elements that can help to better understand the
performance of neural networks.</p>
      <sec id="sec-4-1">
        <title>4.1. Conclusion</title>
        <p>Developing a generic pipeline for processing historical maps will be an important milestone
for massively extracting information from this rich family of cultural heritage documents.
In this research, we made progress in understanding the figurative variance of cartographic
corpora, and established a congruent metric to measure the figurative diversity of a collection
based on the acuity of the multimodal distribution of graphical features. Through several
experiments, we have shown that neural networks are extremely robust in the face of figurative
diversity, even if some map grammars seem more difficult to segment. This high performance
is mostly due to the fact that neural networks can integrate highly abstract reasoning, such
as morphology, topology, and semantic hierarchy, to supplement figurative features, without
excluding the latter. This work, and its conclusions, paves the way for the generic segmentation
of historical maps, highlighting the weaknesses of learning processes and outlining potential
levers for action.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Acknowledgments</title>
      <p>The authors declare to have no conflicts of interest. We would like to thank our former
collaborators, Raphaël Barman and Nils Hamel, for their support on this project.
[37] S. Muhs, H. Herold, G. Meinel, D. Burghardt, and O. Kretschmer. “Automatic
delineation of built-up area at urban block level from topographic maps”. In: Computers,
Environment and Urban Systems 58 (2016), pp. 71–84. doi: 10.1016/j.compenvurbsys.
2016.04.001.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>M. G. Arteaga.</surname>
          </string-name>
          “
          <article-title>Historical map polygon and feature extractor”</article-title>
          .
          <source>In: Proceedings of the 1st ACM SIGSPATIAL International Workshop on MapInteraction. MapInteract '13</source>
          .
          <string-name>
            <surname>Orlando</surname>
          </string-name>
          , Florida: Association for Computing Machinery,
          <year>2013</year>
          , pp.
          <fpage>66</fpage>
          -
          <lpage>71</lpage>
          . doi:
          <volume>10</volume>
          .1145/ 2534931.2534932.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>S.</given-names>
            <surname>Banda</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Agarwal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. R.</given-names>
            <surname>Rao</surname>
          </string-name>
          , and
          <string-name>
            <given-names>R.</given-names>
            <surname>Wankar</surname>
          </string-name>
          . “
          <article-title>Contour layer extraction from colour topographic map by feature selection approach”</article-title>
          .
          <source>In: 2011 IEEE Symposium on Computers Informatics</source>
          .
          <year>2011</year>
          , pp.
          <fpage>425</fpage>
          -
          <lpage>430</lpage>
          . doi:
          <volume>10</volume>
          .1109/isci.
          <year>2011</year>
          .
          <volume>5958953</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>R.</given-names>
            <surname>Barvir</surname>
          </string-name>
          and
          <string-name>
            <given-names>V.</given-names>
            <surname>Vozenilek</surname>
          </string-name>
          . “
          <article-title>Developing Versatile Graphic Map Load Metrics”</article-title>
          .
          <source>In: ISPRS International Journal of Geo-Information</source>
          <volume>9</volume>
          .12 (
          <year>2020</year>
          ), p.
          <fpage>705</fpage>
          . doi:
          <volume>10</volume>
          .3390/ijgi9120705.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>R.</given-names>
            <surname>Brügelmann</surname>
          </string-name>
          . “
          <article-title>Recognition of hatched cartographic patterns”</article-title>
          .
          <source>In: International Archives of Photogrammetry and Remote Sensing</source>
          <volume>31</volume>
          .
          <string-name>
            <surname>B3</surname>
          </string-name>
          (
          <year>1996</year>
          ), pp.
          <fpage>82</fpage>
          -
          <lpage>87</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <surname>Bysyk</surname>
          </string-name>
          . Github bigjpg.
          <year>2019</year>
          . url: https://github.com/by-syk/bigjpg-app.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>Y. S.</given-names>
            <surname>Can</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. J.</given-names>
            <surname>Gerrits</surname>
          </string-name>
          , and
          <string-name>
            <surname>M. E. Kabadayi.</surname>
          </string-name>
          “
          <article-title>Automatic Detection of Road Types From the Third Military Mapping Survey of Austria-Hungary Historical Map Series With Deep Convolutional Neural Networks”</article-title>
          .
          <source>In: IEEE Access 9</source>
          (
          <year>2021</year>
          ), pp.
          <fpage>62847</fpage>
          -
          <lpage>62856</lpage>
          . doi:
          <volume>10</volume>
          .1109/access.
          <year>2021</year>
          .
          <volume>3074897</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>J.</given-names>
            <surname>Chazalon</surname>
          </string-name>
          , E. Carlinet,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Perret</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Duménieu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Mallet</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Géraud</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Nguyen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Nguyen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Baloun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Lenc</surname>
          </string-name>
          , and
          <string-name>
            <given-names>P.</given-names>
            <surname>Král</surname>
          </string-name>
          .
          <source>ICDAR 2021 Competition on Historical Map Segmentation</source>
          .
          <year>2021</year>
          . url: https://arxiv.org/abs/2105.13265.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>Y.-Y.</given-names>
            <surname>Chiang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Duan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Leyk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. H.</given-names>
            <surname>Uhl</surname>
          </string-name>
          , and
          <string-name>
            <given-names>C. A.</given-names>
            <surname>Knoblock</surname>
          </string-name>
          .
          <article-title>Using Historical Maps in Scientific Studies: Applications, Challenges, and Best Practices</article-title>
          . SpringerBriefs in Geography. Cham: Springer International Publishing,
          <year>2020</year>
          . doi:
          <volume>10</volume>
          .1007/978- 3-
          <fpage>319</fpage>
          - 66908-3.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>Y.-Y.</given-names>
            <surname>Chiang</surname>
          </string-name>
          and
          <string-name>
            <given-names>C. A.</given-names>
            <surname>Knoblock</surname>
          </string-name>
          . “
          <article-title>A general approach for extracting road vector data from raster maps”</article-title>
          .
          <source>In: Ijdar 16.1</source>
          (
          <issue>2013</issue>
          ), pp.
          <fpage>55</fpage>
          -
          <lpage>81</lpage>
          . doi:
          <volume>10</volume>
          .1007/s10032-011-0177-1.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>Y.-Y.</given-names>
            <surname>Chiang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Leyk</surname>
          </string-name>
          , and
          <string-name>
            <given-names>C. A.</given-names>
            <surname>Knoblock</surname>
          </string-name>
          .
          <article-title>“A Survey of Digital Map Processing Techniques”</article-title>
          .
          <source>In: ACM Comput. Surv. 47.1</source>
          (
          <issue>2014</issue>
          ),
          <volume>1</volume>
          :
          <fpage>1</fpage>
          -
          <lpage>1</lpage>
          :
          <fpage>44</fpage>
          . doi:
          <volume>10</volume>
          .1145/2557423.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>Y.-Y.</given-names>
            <surname>Chiang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Leyk</surname>
          </string-name>
          , and
          <string-name>
            <given-names>C. A.</given-names>
            <surname>Knoblock</surname>
          </string-name>
          . “
          <article-title>Efficient and Robust Graphics Recognition from Historical Maps”</article-title>
          .
          <source>In: Graphics Recognition. New Trends and Challenges</source>
          . Ed. by Y.-
          <string-name>
            <given-names>B.</given-names>
            <surname>Kwon</surname>
          </string-name>
          and
          <string-name>
            <surname>J.-M. Ogier</surname>
          </string-name>
          . Lecture Notes in Computer Science. Berlin, Heidelberg: Springer,
          <year>2013</year>
          , pp.
          <fpage>25</fpage>
          -
          <lpage>35</lpage>
          . doi:
          <volume>10</volume>
          .1007/978-3-
          <fpage>642</fpage>
          -36824-0\_3.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>A.</given-names>
            <surname>Cordeiro</surname>
          </string-name>
          and
          <string-name>
            <given-names>P.</given-names>
            <surname>Pina</surname>
          </string-name>
          . “
          <article-title>Colour map object separation”</article-title>
          . In: Remote Sensing: From Pixels to Processes.
          <year>2006</year>
          , pp.
          <fpage>243</fpage>
          -
          <lpage>247</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>N.</given-names>
            <surname>Dalal</surname>
          </string-name>
          and
          <string-name>
            <given-names>B.</given-names>
            <surname>Triggs</surname>
          </string-name>
          . “
          <article-title>Histograms of Oriented Gradients for Human Detection”</article-title>
          .
          <source>In: vol. 1</source>
          . IEEE Computer Society,
          <year>2005</year>
          , pp.
          <fpage>886</fpage>
          -
          <lpage>893</lpage>
          . doi:
          <volume>10</volume>
          .1109/cvpr.
          <year>2005</year>
          .
          <volume>177</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>J.</given-names>
            <surname>Deng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Dong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Socher</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.-J.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Li</surname>
          </string-name>
          , and
          <string-name>
            <given-names>L.</given-names>
            <surname>Fei-Fei</surname>
          </string-name>
          .
          <article-title>“ImageNet: A large-scale hierarchical image database”</article-title>
          .
          <source>In: IEEE Conference on Computer Vision and Pattern Recognition</source>
          .
          <year>2009</year>
          , pp.
          <fpage>248</fpage>
          -
          <lpage>255</lpage>
          . doi:
          <volume>10</volume>
          .1109/cvpr.
          <year>2009</year>
          .
          <volume>5206848</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>D. B.</given-names>
            <surname>Dhar</surname>
          </string-name>
          and
          <string-name>
            <given-names>B.</given-names>
            <surname>Chanda</surname>
          </string-name>
          . “
          <article-title>Extraction and recognition of geographical features from paper maps”</article-title>
          .
          <source>In: Ijdar 8.4</source>
          (
          <issue>2006</issue>
          ), pp.
          <fpage>232</fpage>
          -
          <lpage>245</lpage>
          . doi:
          <volume>10</volume>
          .1007/s10032-005-0010-9.
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>H.</given-names>
            <surname>Ernstson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. E. van der</given-names>
            <surname>Leeuw</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. L.</given-names>
            <surname>Redman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. J.</given-names>
            <surname>Mefert</surname>
          </string-name>
          , G. Davis,
          <string-name>
            <given-names>C.</given-names>
            <surname>Alfsen</surname>
          </string-name>
          , and
          <string-name>
            <given-names>T.</given-names>
            <surname>Elmqvist</surname>
          </string-name>
          . “Urban Transitions:
          <article-title>On Urban Resilience and Human-Dominated Ecosystems”</article-title>
          .
          <source>In: Ambio 39.8</source>
          (
          <issue>2010</issue>
          ), pp.
          <fpage>531</fpage>
          -
          <lpage>545</lpage>
          . doi:
          <volume>10</volume>
          .1007/s13280-010-0081-9.
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>A.</given-names>
            <surname>Garcia-Molsosa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H. A.</given-names>
            <surname>Orengo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Lawrence</surname>
          </string-name>
          , G. Philip,
          <string-name>
            <given-names>K.</given-names>
            <surname>Hopper</surname>
          </string-name>
          , and
          <string-name>
            <given-names>C. A.</given-names>
            <surname>Petrie</surname>
          </string-name>
          . “
          <article-title>Potential of deep learning segmentation for the extraction of archaeological features from historical map series”</article-title>
          .
          <source>In: Archaeological Prospection 28.2</source>
          (
          <issue>2021</issue>
          ), pp.
          <fpage>187</fpage>
          -
          <lpage>199</lpage>
          . doi:
          <volume>10</volume>
          .1002/arp.
          <year>1807</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>X.</given-names>
            <surname>Glorot</surname>
          </string-name>
          and
          <string-name>
            <given-names>Y.</given-names>
            <surname>Bengio</surname>
          </string-name>
          . “
          <article-title>Understanding the difficulty of training deep feedforward neural networks”</article-title>
          .
          <source>In: Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics</source>
          . Ed. by
          <string-name>
            <given-names>Y. W.</given-names>
            <surname>Teh</surname>
          </string-name>
          and
          <string-name>
            <given-names>M.</given-names>
            <surname>Titterington</surname>
          </string-name>
          . Vol.
          <volume>9</volume>
          .
          <source>Proceedings of Machine Learning Research. Chia Laguna Resort</source>
          , Sardinia, Italy: Pmlr,
          <year>2010</year>
          , pp.
          <fpage>249</fpage>
          -
          <lpage>256</lpage>
          . url: http://proceedings.mlr.press/v9/glorot10a.html.
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>B.</given-names>
            <surname>Graef</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Carosio</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Graef</surname>
          </string-name>
          , and
          <string-name>
            <given-names>R.</given-names>
            <surname>Carosio</surname>
          </string-name>
          . “
          <article-title>Automatic Interpretation of RasterBased Topographic Maps by Means of Queries”</article-title>
          . In: FIG XXII International Congress Washington, D. C.,
          <string-name>
            <surname>published on</surname>
            <given-names>CD-ROM</given-names>
          </string-name>
          (
          <year>2002</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>K.</given-names>
            <surname>He</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , S. Ren, and
          <string-name>
            <given-names>J.</given-names>
            <surname>Sun</surname>
          </string-name>
          .
          <article-title>Deep Residual Learning for Image Recognition</article-title>
          .
          <year>2015</year>
          . url: https://arxiv.org/abs/1512.03385.
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>M.</given-names>
            <surname>Heitzler</surname>
          </string-name>
          and
          <string-name>
            <given-names>L.</given-names>
            <surname>Hurni</surname>
          </string-name>
          . “
          <article-title>Cartographic reconstruction of building footprints from historical maps: A study on the Swiss Siegfried map”</article-title>
          .
          <source>In: Transactions in GIS 24.2</source>
          (
          <issue>2020</issue>
          ), pp.
          <fpage>442</fpage>
          -
          <lpage>461</lpage>
          . doi:
          <volume>10</volume>
          .1111/tgis.12610.
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>K.</given-names>
            <surname>Hosseini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>McDonough</surname>
          </string-name>
          ,
          <string-name>
            <surname>D. van Strien</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Vane</surname>
          </string-name>
          , and
          <string-name>
            <given-names>D. C. S.</given-names>
            <surname>Wilson</surname>
          </string-name>
          . “
          <article-title>Maps of a Nation? The Digitized Ordnance Survey for New Historical Research”</article-title>
          .
          <source>In: Journal of Victorian Culture 26.2</source>
          (
          <issue>2021</issue>
          ), pp.
          <fpage>284</fpage>
          -
          <lpage>299</lpage>
          . doi:
          <volume>10</volume>
          .1093/jvcult/vcab009.
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>G.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. Van Der</given-names>
            <surname>Maaten</surname>
          </string-name>
          , and
          <string-name>
            <given-names>K. Q.</given-names>
            <surname>Weinberger</surname>
          </string-name>
          . “
          <article-title>Densely Connected Convolutional Networks”</article-title>
          .
          <source>In: 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)</source>
          .
          <year>2017</year>
          , pp.
          <fpage>2261</fpage>
          -
          <lpage>2269</lpage>
          . doi:
          <volume>10</volume>
          .1109/cvpr.
          <year>2017</year>
          .
          <volume>243</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>A.</given-names>
            <surname>Karmilof-Smith.</surname>
          </string-name>
          <article-title>“Constraints on representational change: evidence from children's drawing”</article-title>
          .
          <source>In: Cognition 34.1</source>
          (
          <issue>1990</issue>
          ), pp.
          <fpage>57</fpage>
          -
          <lpage>83</lpage>
          . doi:
          <volume>10</volume>
          .1016/
          <fpage>0010</fpage>
          -
          <lpage>0277</lpage>
          (
          <issue>90</issue>
          )
          <fpage>90031</fpage>
          -e.
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>E.</given-names>
            <surname>Katona</surname>
          </string-name>
          and
          <string-name>
            <surname>G. Hudra. “</surname>
          </string-name>
          <article-title>An interpretation system for cadastral maps”</article-title>
          .
          <source>In: Proceedings 10th International Conference on Image Analysis and Processing</source>
          .
          <year>1999</year>
          , pp.
          <fpage>792</fpage>
          -
          <lpage>797</lpage>
          . doi:
          <volume>10</volume>
          .1109/iciap.
          <year>1999</year>
          .
          <volume>797692</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [26]
          <string-name>
            <given-names>N. W.</given-names>
            <surname>Kim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>and J.</given-names>
            <surname>Seo</surname>
          </string-name>
          . “
          <article-title>Accurate segmentation of land regions in historical cadastral maps”</article-title>
          .
          <source>In: Journal of Visual Communication and Image Representation 25.5</source>
          (
          <issue>2014</issue>
          ), pp.
          <fpage>1262</fpage>
          -
          <lpage>1274</lpage>
          . doi:
          <volume>10</volume>
          .1016/j.jvcir.
          <year>2014</year>
          .
          <volume>01</volume>
          .001.
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [27]
          <string-name>
            <given-names>D. H.</given-names>
            <surname>Laboratory.</surname>
          </string-name>
          dhSegment-torch.
          <year>2021</year>
          . url: https : / / github . com / dhlab - epfl / dhSegment-torch.
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          [28]
          <string-name>
            <given-names>R. G.</given-names>
            <surname>Laycock</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Drinkwater</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <given-names>A. M.</given-names>
            <surname>Day</surname>
          </string-name>
          . “
          <article-title>Exploring cultural heritage sites through space and time”</article-title>
          .
          <source>In: J. Comput. Cult. Herit. 1</source>
          .
          <issue>2</issue>
          (
          <issue>2008</issue>
          ),
          <volume>11</volume>
          :
          <fpage>1</fpage>
          -
          <lpage>11</lpage>
          :
          <fpage>15</fpage>
          . doi:
          <volume>10</volume>
          .1145/1434763. 1434768.
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          [29]
          <string-name>
            <given-names>S. D.</given-names>
            <surname>Laycock</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. G.</given-names>
            <surname>Brown</surname>
          </string-name>
          , R. G.
          <article-title>Laycock, and</article-title>
          <string-name>
            <given-names>A. M.</given-names>
            <surname>Day</surname>
          </string-name>
          . “
          <article-title>Aligning archive maps and extracting footprints for analysis of historic urban environments”</article-title>
          .
          <source>In: Computers &amp; Graphics. Virtual Reality in Brazil 35.2</source>
          (
          <issue>2011</issue>
          ), pp.
          <fpage>242</fpage>
          -
          <lpage>249</lpage>
          . doi:
          <volume>10</volume>
          .1016/j.cag.
          <year>2011</year>
          .
          <volume>01</volume>
          . 002.
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          [30]
          <string-name>
            <given-names>S.</given-names>
            <surname>Leyk</surname>
          </string-name>
          . “
          <article-title>Segmentation of Colour Layers in Historical Maps Based on Hierarchical Colour Sampling”</article-title>
          .
          <source>In: Graphics Recognition. Achievements</source>
          , Challenges, and Evolution. Ed. by
          <string-name>
            <surname>J.-M. Ogier</surname>
          </string-name>
          , W. Liu, and J.
          <source>Lladós. Lecture Notes in Computer Science</source>
          . Berlin, Heidelberg: Springer,
          <year>2010</year>
          , pp.
          <fpage>231</fpage>
          -
          <lpage>241</lpage>
          . doi:
          <volume>10</volume>
          .1007/978-3-
          <fpage>642</fpage>
          -13728-0\_
          <fpage>21</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          [31]
          <string-name>
            <given-names>S.</given-names>
            <surname>Leyk</surname>
          </string-name>
          and
          <string-name>
            <given-names>R.</given-names>
            <surname>Boesch</surname>
          </string-name>
          . “
          <article-title>Colors of the past: color image segmentation in historical topographic maps based on homogeneity”</article-title>
          .
          <source>In: Geoinformatica 14.1</source>
          (
          <issue>2009</issue>
          ), p.
          <fpage>1</fpage>
          . doi:
          <volume>10</volume>
          .1007/s10707-008-0074-z.
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          [32]
          <string-name>
            <given-names>J.</given-names>
            <surname>Long</surname>
          </string-name>
          , E. Shelhamer, and
          <string-name>
            <given-names>T.</given-names>
            <surname>Darrell</surname>
          </string-name>
          . “
          <article-title>Fully Convolutional Networks for Semantic Segmentation”</article-title>
          .
          <source>In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition</source>
          .
          <year>2015</year>
          , pp.
          <fpage>3431</fpage>
          -
          <lpage>3440</lpage>
          . url: https://arxiv.org/abs/1411.4038.
        </mixed-citation>
      </ref>
      <ref id="ref33">
        <mixed-citation>
          [33]
          <string-name>
            <given-names>C.</given-names>
            <surname>Mello</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Costa</surname>
          </string-name>
          , and T. d. Santos. “
          <article-title>Automatic image segmentation of old topographic maps and floor plans”</article-title>
          .
          <source>In: 2012 IEEE International Conference on Systems, Man, and Cybernetics (SMC)</source>
          .
          <year>2012</year>
          , pp.
          <fpage>132</fpage>
          -
          <lpage>137</lpage>
          . doi:
          <volume>10</volume>
          .1109/icsmc.
          <year>2012</year>
          .
          <volume>6377689</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref34">
        <mixed-citation>
          [34]
          <string-name>
            <given-names>Q.</given-names>
            <surname>Miao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Xu</surname>
          </string-name>
          , T. Liu,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , and
          <string-name>
            <given-names>W.</given-names>
            <surname>Li</surname>
          </string-name>
          . “
          <article-title>Linear Feature Separation From Topographic Maps Using Energy Density and the Shear Transform”</article-title>
          .
          <source>In: IEEE Transactions on Image Processing 22.4</source>
          (
          <issue>2013</issue>
          ), pp.
          <fpage>1548</fpage>
          -
          <lpage>1558</lpage>
          . doi:
          <volume>10</volume>
          .1109/tip.
          <year>2012</year>
          .
          <volume>2233487</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref35">
        <mixed-citation>
          [35]
          <string-name>
            <given-names>T.</given-names>
            <surname>Miyoshi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Kaneda</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Yamashita</surname>
          </string-name>
          , and
          <string-name>
            <surname>E. Nakamae.</surname>
          </string-name>
          “
          <article-title>Automatic extraction of buildings utilizing geometric features of a scanned topographic map”</article-title>
          .
          <source>In: Proceedings of the 17th International Conference on Pattern Recognition</source>
          ,
          <year>2004</year>
          .
          <source>ICPR 2004</source>
          . Vol.
          <volume>3</volume>
          .
          <year>2004</year>
          ,
          <fpage>626</fpage>
          -
          <lpage>629</lpage>
          Vol.
          <volume>3</volume>
          . doi:
          <volume>10</volume>
          .1109/icpr.
          <year>2004</year>
          .
          <volume>1334607</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref36">
        <mixed-citation>
          [36]
          <string-name>
            <given-names>S.</given-names>
            <surname>Muhs</surname>
          </string-name>
          . “
          <article-title>Computational Delineation of Built-up Area at Urban Block Level from Topographic Maps: A Contribution to Retrospective Monitoring of Urban Dynamics”</article-title>
          .
          <source>PhD thesis</source>
          . Dresden, Germany: Technische Universität Dresden,
          <year>2019</year>
          . url: https : / / nbn - resolving.org/urn:nbn:de:bsz:
          <fpage>14</fpage>
          -
          <lpage>qucosa2</lpage>
          -340364.
        </mixed-citation>
      </ref>
      <ref id="ref37">
        <mixed-citation>
          [38]
          <string-name>
            <given-names>S.</given-names>
            <surname>Muhs</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Meinel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Burghardt</surname>
          </string-name>
          , and
          <string-name>
            <given-names>H.</given-names>
            <surname>Herold</surname>
          </string-name>
          . “
          <article-title>Automatisierte Baublockabgrenzung in Topographischen Karten”</article-title>
          . In: Flächennutzungsmonitoring. ...
          <string-name>
            <surname>Flächennutzungsmonitoring</surname>
            <given-names>V</given-names>
          </string-name>
          : Methodik, Analyseergebnisse, Flächenmanagement.
          <source>IÖR Schriften 61</source>
          . Berlin: Rhombos-Verl,
          <year>2013</year>
          , pp.
          <fpage>211</fpage>
          -
          <lpage>219</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref38">
        <mixed-citation>
          [39]
          <string-name>
            <surname>J.-M. Ogier</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Mullot</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Labiche</surname>
            , and
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>Lecourtier</surname>
          </string-name>
          . “
          <article-title>Technical Map Interpretation: A Distributed Approach”</article-title>
          .
          <source>In: Pattern Analysis &amp; Applications 3</source>
          .2 (
          <issue>2000</issue>
          ), pp.
          <fpage>88</fpage>
          -
          <lpage>103</lpage>
          . doi:
          <volume>10</volume>
          .1007/pl00010983.
        </mixed-citation>
      </ref>
      <ref id="ref39">
        <mixed-citation>
          [40]
          <string-name>
            <given-names>T.</given-names>
            <surname>Ojala</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Pietikainen</surname>
          </string-name>
          , and
          <string-name>
            <given-names>T.</given-names>
            <surname>Maenpaa</surname>
          </string-name>
          . “
          <article-title>Multiresolution gray-scale and rotation invariant texture classification with local binary patterns”</article-title>
          .
          <source>In: IEEE Transactions on Pattern Analysis and Machine Intelligence</source>
          <volume>24</volume>
          .7 (
          <issue>2002</issue>
          ), pp.
          <fpage>971</fpage>
          -
          <lpage>987</lpage>
          . doi:
          <volume>10</volume>
          . 1109 / tpami.
          <year>2002</year>
          .
          <volume>1017623</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref40">
        <mixed-citation>
          [41]
          <string-name>
            <given-names>S.</given-names>
            <surname>Oliveira</surname>
          </string-name>
          , I. di Lenardo,
          <string-name>
            <given-names>B.</given-names>
            <surname>Tourenc</surname>
          </string-name>
          , and
          <string-name>
            <given-names>F.</given-names>
            <surname>Kaplan</surname>
          </string-name>
          .
          <article-title>“A deep learning approach to Cadastral Computing”</article-title>
          . In: Utrecht, Netherlands,
          <year>2019</year>
          . url: https : / / dev . clariah . nl / files/dh2019/boa/0691.html.
        </mixed-citation>
      </ref>
      <ref id="ref41">
        <mixed-citation>
          [42]
          <string-name>
            <given-names>S. A.</given-names>
            <surname>Oliveira</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Seguin</surname>
          </string-name>
          , and
          <string-name>
            <given-names>F.</given-names>
            <surname>Kaplan</surname>
          </string-name>
          . “dhSegment:
          <article-title>A generic deep-learning approach for document segmentation”</article-title>
          .
          <source>In: 16th International Conference on Frontiers in Handwriting Recognition (ICFHR)</source>
          (
          <year>2018</year>
          ), pp.
          <fpage>7</fpage>
          -
          <lpage>12</lpage>
          . doi:
          <volume>10</volume>
          .1109/icfhr-2018.
          <year>2018</year>
          .
          <volume>00011</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref42">
        <mixed-citation>
          [43]
          <string-name>
            <given-names>N.</given-names>
            <surname>Otsu</surname>
          </string-name>
          .
          <article-title>“A Threshold Selection Method from Gray-Level Histograms”</article-title>
          .
          <source>In: IEEE Transactions on Systems, Man, and Cybernetics 9</source>
          .1 (
          <issue>1979</issue>
          ), pp.
          <fpage>62</fpage>
          -
          <lpage>66</lpage>
          . doi:
          <volume>10</volume>
          .1109/tsmc.
          <year>1979</year>
          .
          <volume>4310076</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref43">
        <mixed-citation>
          [44]
          <string-name>
            <given-names>R.</given-names>
            <surname>Petitpierre</surname>
          </string-name>
          . Generic Semantic Segmentation of Historical Maps - Github repository.
          <year>2021</year>
          . url: https://github.com/RPetitpierre/Generic%5C
          <source>%5FSemantic%5C% 5FSegmentation%5C%5Fof%5C%5FHistorical%5C%5FMaps.</source>
        </mixed-citation>
      </ref>
      <ref id="ref44">
        <mixed-citation>
          [45]
          <string-name>
            <given-names>R.</given-names>
            <surname>Petitpierre</surname>
          </string-name>
          .
          <source>Historical City Maps Semantic Segmentation Dataset</source>
          .
          <year>2021</year>
          . doi:
          <volume>10</volume>
          .5281/ zenodo.5513639.
        </mixed-citation>
      </ref>
      <ref id="ref45">
        <mixed-citation>
          [46]
          <string-name>
            <given-names>R.</given-names>
            <surname>Petitpierre</surname>
          </string-name>
          . “
          <article-title>Neural networks for semantic segmentation of historical city maps: Crosscultural performance and the impact of figurative diversity”</article-title>
          .
          <source>In: CoRR abs/2101</source>
          .12478 (
          <year>2021</year>
          ). url: https://arxiv.org/abs/2101.12478.
        </mixed-citation>
      </ref>
      <ref id="ref46">
        <mixed-citation>
          [47] “
          <article-title>Procédé d'impression des cartes géographiques en couleur”</article-title>
          . In: L'
          <article-title>Echo du monde savant</article-title>
          ,
          <source>journal analytique des nouvelles et des cours scientifiques . 2</source>
          . Paris,
          <year>1840</year>
          , pp.
          <fpage>401</fpage>
          -
          <lpage>402</lpage>
          . url: https://go.epfl.
          <source>ch/procede%5C%5Fdimpression%5C%5F1840.</source>
        </mixed-citation>
      </ref>
      <ref id="ref47">
        <mixed-citation>
          [48]
          <string-name>
            <given-names>O.</given-names>
            <surname>Ronneberger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Fischer</surname>
          </string-name>
          , and
          <string-name>
            <given-names>T.</given-names>
            <surname>Brox</surname>
          </string-name>
          . “U-Net:
          <article-title>Convolutional Networks for Biomedical Image Segmentation”</article-title>
          . In: arXiv:
          <fpage>1505</fpage>
          .04597 [cs] (
          <year>2015</year>
          ). url: http://arxiv.org/abs/1505. 04597.
        </mixed-citation>
      </ref>
      <ref id="ref48">
        <mixed-citation>
          [49]
          <string-name>
            <given-names>S.</given-names>
            <surname>Salvatore</surname>
          </string-name>
          and
          <string-name>
            <given-names>P.</given-names>
            <surname>Guitton</surname>
          </string-name>
          . “
          <article-title>Contour line recognition from scanned topographic maps”</article-title>
          . In: (
          <year>2004</year>
          ). url: http://dspace5.zcu.cz/handle/11025/1744.
        </mixed-citation>
      </ref>
      <ref id="ref49">
        <mixed-citation>
          [50]
          <string-name>
            <given-names>A.</given-names>
            <surname>Savitzky</surname>
          </string-name>
          and
          <string-name>
            <surname>M. J. E. Golay.</surname>
          </string-name>
          “
          <article-title>Smoothing and Diferentiation of Data by Simplified Least Squares Procedures</article-title>
          .”
          <source>In: Anal. Chem</source>
          .
          <volume>36</volume>
          .8 (
          <issue>1964</issue>
          ), pp.
          <fpage>1627</fpage>
          -
          <lpage>1639</lpage>
          . doi:
          <volume>10</volume>
          .1021/ ac60214a047.
        </mixed-citation>
      </ref>
      <ref id="ref50">
        <mixed-citation>
          [51]
          <string-name>
            <given-names>D.</given-names>
            <surname>Schemala</surname>
          </string-name>
          . “
          <article-title>Semantische Segmentierung historischer topographischer Karten”</article-title>
          .
          <source>PhD thesis</source>
          . Dresden, Germany: Technische Universität Dresden,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref51">
        <mixed-citation>
          [52]
          <string-name>
            <given-names>G.</given-names>
            <surname>Touya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Decherf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Lalanne</surname>
          </string-name>
          , and
          <string-name>
            <given-names>M.</given-names>
            <surname>Dumont</surname>
          </string-name>
          . “
          <article-title>Comparing image-based methods for assessing visual clutter in generalized maps”</article-title>
          .
          <source>In: ISPRS Annals of the Photogrammetry, Remote Sensing and Spatial Information Sciences Ii-3/w5</source>
          (
          <year>2015</year>
          ), pp.
          <fpage>227</fpage>
          -
          <lpage>233</lpage>
          . doi:
          <volume>10</volume>
          .5194/isprsannals-II-3
          <string-name>
            <surname>-W5-</surname>
          </string-name>
          227-
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref52">
        <mixed-citation>
          [53]
          <string-name>
            <given-names>J. H.</given-names>
            <surname>Uhl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Leyk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.-Y.</given-names>
            <surname>Chiang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Duan</surname>
          </string-name>
          , and
          <string-name>
            <given-names>C. A.</given-names>
            <surname>Knoblock</surname>
          </string-name>
          . “
          <source>Automated Extraction of Human Settlement Patterns From Historical Topographic Map Series Using Weakly Supervised Convolutional Neural Networks”. In: IEEE Access 8</source>
          (
          <year>2020</year>
          ), pp.
          <fpage>6978</fpage>
          -
          <lpage>6996</lpage>
          . doi:
          <volume>10</volume>
          .1109/access.
          <year>2019</year>
          .
          <volume>2963213</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref53">
        <mixed-citation>
          [54]
          <string-name>
            <given-names>J. H.</given-names>
            <surname>Uhl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Leyk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.-Y.</given-names>
            <surname>Chiang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Duan</surname>
          </string-name>
          , and
          <string-name>
            <given-names>C. A.</given-names>
            <surname>Knoblock</surname>
          </string-name>
          . “Map Archive Mining:
          <article-title>Visual-Analytical Approaches to Explore Large Historical Map Collections”</article-title>
          .
          <source>In: ISPRS International Journal of Geo-Information 7.4</source>
          (
          <issue>2018</issue>
          ), p.
          <fpage>148</fpage>
          . doi:
          <volume>10</volume>
          .3390/ijgi7040148.
        </mixed-citation>
      </ref>
      <ref id="ref54">
        <mixed-citation>
          [55]
          <string-name>
            <surname>J.-M. Viglino</surname>
            and
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Pierrot-Deseilligny</surname>
          </string-name>
          .
          <article-title>“A vector approach for automatic interpretation of the French cadastral map”</article-title>
          .
          <source>In: 7th International Conference on Document Analysis and Recognition. Proceedings</source>
          .
          <year>2003</year>
          ,
          <fpage>304</fpage>
          -
          <lpage>308</lpage>
          vol.
          <volume>1</volume>
          . doi:
          <volume>10</volume>
          .1109/icdar.
          <year>2003</year>
          .
          <volume>1227678</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref55">
        <mixed-citation>
          [56]
          <string-name>
            <surname>J. Wu</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Wei</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          <string-name>
            <surname>Yuan</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          <string-name>
            <surname>Shu</surname>
            ,
            <given-names>Y.-Y.</given-names>
          </string-name>
          <string-name>
            <surname>Chiang</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          <string-name>
            <surname>Fu</surname>
            , and
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Deng</surname>
          </string-name>
          .
          <article-title>“A New Gabor FilterBased Method for Automatic Recognition of Hatched Residential Areas”</article-title>
          .
          <source>In: IEEE Access 7</source>
          (
          <year>2019</year>
          ), pp.
          <fpage>40649</fpage>
          -
          <lpage>40662</lpage>
          . doi:
          <volume>10</volume>
          .1109/access.
          <year>2019</year>
          .
          <volume>2907114</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref56">
        <mixed-citation>
          [57]
          <string-name>
            <given-names>H.</given-names>
            <surname>Yamada</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Yamamoto</surname>
          </string-name>
          , and
          <string-name>
            <given-names>K.</given-names>
            <surname>Hosokawa</surname>
          </string-name>
          . “
          <article-title>Directional mathematical morphology and reformalized Hough transformation for the analysis of topographic maps”</article-title>
          .
          <source>In: IEEE Transactions on Pattern Analysis and Machine Intelligence</source>
          <volume>15</volume>
          .4 (
          <issue>1993</issue>
          ), pp.
          <fpage>380</fpage>
          -
          <lpage>387</lpage>
          . doi:
          <volume>10</volume>
          .1109/34.206957.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>