<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>Binary Codes Capable of Correct-
ing Deletions, Insertions and Reversals. Soviet Physics
Doklady</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Noe¨mi Aepli</string-name>
          <email>noemi.aepli@uzh.ch</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Simon Clematide</string-name>
          <email>simon.clematide@cl.uzh.ch</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>University of Zurich</institution>
        </aff>
      </contrib-group>
      <pub-date>
        <year>1966</year>
      </pub-date>
      <volume>10</volume>
      <issue>707</issue>
      <abstract>
        <p>This paper presents different approaches towards universal dependency parsing for Swiss German. Dealing with dialects is a challenging task in Natural Language Processing because of the huge linguistic variability, which is partly due to the lack of standard spelling rules. Building a statistical parser requires expensive resources which are only available for a few dozen high-resourced languages. In order to overcome the lowresource problem for dialects, approaches to cross-lingual learning are exploited. We apply different cross-lingual parsing strategies to Swiss German, making use of Standard German resources. The methods applied are annotation projection and model transfer. The results show around 60% Labelled Attachment Score for all approaches and provide a first substantial step towards Swiss German dependency parsing. The resources are available for further research on NLP applications for Swiss German dialects.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1 Introduction</title>
      <p>Swiss German is a dialect continuum of the
Alemannic dialect group, comprising numerous varieties used
in the German-speaking part of Switzerland.1 Unlike
other dialect situations, the Swiss German dialects are
deeply rooted in the Swiss culture and enjoy a high
reputation, i.e. dialect speakers are not considered less
educated as it is the case in other countries. On the
basis of their high acceptance in the Swiss culture and
with the introduction of digital communication, Swiss
German has undergone a spread over all kinds of
communication forms and social media. Despite being
oral languages, the dialects are used increasingly in
written contexts, and writers spell as they please.</p>
      <p>For Natural Language Processing (NLP),
lowresourced languages are challenging, particularly in
cases like Swiss German where no orthographic rules
are followed. Compiling NLP resources from scratch
such as syntactically annotated text corpora
(treebanks) is a laborious and expensive process. Thus, in
such cases, cross-lingual approaches offer a
perspective to get started with automatic processing of the
respective language. Such approaches are especially
promising if a closely related resource-rich language
is available, which is the case for Swiss German.</p>
      <p>The Universal Dependencies (UD) project aims
at developing and setting a standard for
crosslinguistically consistently annotated treebanks in
order to facilitate multilingual parsing research. We
support this idea by adopting the current UD standard as
much as possible.</p>
      <p>The information about which word of the sentence
is dependent on which other one is important in
order to correctly understand the meaning of a sentence.
Thus, it is needed for numerous NLP applications like
information extraction or grammar checking. The task
of identifying these dependencies is done by a
dependency parser (see Figure 1 for a Swiss German
example in UD).</p>
      <p>In this paper, we apply two different cross-lingual
dependency parsing strategies, namely annotation
projection as a lexicalised approach, and model
transfer as a delexicalised approach. We manually create a
gold standard in order to evaluate and compare the
different strategies. Furthermore, we build and evaluate</p>
    </sec>
    <sec id="sec-2">
      <title>Related Work</title>
      <p>
        Even though there have been several projects
involving Swiss German
        <xref ref-type="bibr" rid="ref1 ref10 ref13 ref15 ref19 ref23 ref24 ref25 ref26 ref27 ref28 ref3 ref31 ref31 ref36 ref38 ref8">(Hollenstein and Aepli, 2014;
Zampieri et al., 2017; Hollenstein and Aepli, 2015;
Samardzic et al., 2016; Samardzˇic´ et al., 2015;
Scherrer, 2007; Baumgartner, 2016; Du¨rscheid and Stark,
2011; Stark et al., 2014; Scherrer and Owen, 2010;
Scherrer, 2013, 2012)</xref>
        , resources for NLP applications
are still rare. As so often for dialects, even data for
Swiss German is sparse. Therefore, the approach is to
use tools and data of related resource-rich languages
and apply transfer methods.
2.1
      </p>
      <sec id="sec-2-1">
        <title>Universal Dependencies</title>
        <p>
          Research in dependency parsing has increased
significantly since a collection of dependency treebanks has
become available, in particular through the CoNLL
shared tasks on dependency parsing
          <xref ref-type="bibr" rid="ref17 ref18 ref22 ref4 ref40 ref6">(Buchholz and
Marsi, 2006; Nivre et al., 2007a; Zeman et al., 2017)</xref>
          which have provided many data sets. In order to
facilitate cross-lingual research on syntactic structure and
to standardise best-practices, Universal POS (UPOS)
tags
          <xref ref-type="bibr" rid="ref20">(Petrov et al., 2012)</xref>
          as well as Universal
Dependencies
          <xref ref-type="bibr" rid="ref16">(Nivre et al., 2016)</xref>
          have been introduced. The
annotation scheme is originally based on Stanford
dependencies (de Marneffe et al., 2006; de Marneffe and
        </p>
        <p>
          There are two main approaches to cross-lingual
syntactic dependency parsing. Firstly, the delexicalized
model transfer of which the goal is to abstract away
from language-specific parameters, i.e. train
delexicalised parsers. The idea is based on universal
features and model parameters that can be transferred
between related languages. Hence, this method assumes
a common feature representation across languages.
The advantage of the model transfer approach is that
no parallel data is nee
          <xref ref-type="bibr" rid="ref39">ded. Zeman and Resnik (2008</xref>
          )
train a basic delexicalised parser relying on
part-ofspeech (POS) tags only. McDonald et al. (2013);
Petrov et al. (2012) and Naseem et al. (2010) rely
on universal features while Ta¨ckstro¨m et al. (2013)
adapt model parameters to the target language in order
to cross-linguistically transfer syntactic dependency
parses.
        </p>
        <p>
          The main idea of the second approach, the
lexicalised annotation projection method, is the mapping
of labels across languages using parallel sentences and
automatic alignment. It includes projection heuristics
and usually post projection rules. The main drawback
of this approach is that it relies on sentence-aligned
parallel corpora. In order to deal with this restriction,
treebank translation has emerged where the training
data is automatically translated with a machine
translation system. The central point of this method is the
alignment along which the annotations are mapped
from one language to the other. Automatic word
alignment has already been used by Yarowsky et al.
(2001); Aepli et al. (2014) an
          <xref ref-type="bibr" rid="ref39">d Snyder et al. (2008</xref>
          )
for improving resources and tools for POS tagging
of supervised and unsupervised learning respectively.
Hwa et al. (2005), Tiedemann (2014) and Tiedemann
(2015) use annotation projection approaches for
parsing, and Tiedemann et al. (2014) as well as Rosa et al.
(2017) use machine translation in addition instead of
relying on parallel corpora. For Swiss German,
treebank translation is not viable because of sparse data
and the lack of a Machine Translation system for
Swiss German. Hence, in this paper we apply
annotation projection as a lexicalised approach and model
transfer as a delexicalised approach.
3
3.1
        </p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Materials</title>
      <sec id="sec-3-1">
        <title>Standard German Data</title>
        <p>
          We use the German Universal Dependency treebank2
consisting of 13,814 sentences. It is annotated
according to the UD guidelines3 and contains
Universal POS (UPOS) tags
          <xref ref-type="bibr" rid="ref20">(Petrov et al., 2012)</xref>
          . The
treebank comes in CoNLL-U format but as some tools
cannot handle it, we convert it to CoNLL-X. This
includes one major tokenization change concerning
the Stuttgart-Tu¨bingen-TagSet (STTS)
          <xref ref-type="bibr" rid="ref29">(Schiller et al.,
1999)</xref>
          POS tag APPRART. In CoNLL-U the
prepositions with fused articles are split into two syntactical
words. We undo this split, merge the information in
one token and correspondingly adapt the dependency
relations.
3.2
        </p>
      </sec>
      <sec id="sec-3-2">
        <title>Swiss German Data</title>
        <p>
          Annotation projection requires a parallel corpus.
The AGORA citizen linguistics project4 crowdsourced
Standard German translations of 6,197 Swiss German
sentences via the web site dindialaekt.ch. The
sentences are taken from the NOAH corpus
          <xref ref-type="bibr" rid="ref1 ref10 ref31 ref36">(Hollenstein
and Aepli, 2014)</xref>
          , additionally, sentences from novels
in Bernese and St Gallen dialect were added to
better represent syntactic word order differences. By the
end of November 2017, the citizen linguists produced
41,670 translations. We aggregated and cleaned the
data into a parallel GSW/DE corpus of 26,015
sentences. In particular, we filtered translations that
dif2https://github.com/UniversalDependencies/UD German
3http://universaldependencies.org/guidelines.html
4https://www.linguistik.uzh.ch/de/forschung/agora.html
fered too much in length or Levenshtein edit distance5
from the Swiss German source sentence.
4
        </p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Methods</title>
      <p>
        We apply two classical parsing approaches presented
in Section 2: model transfer with a delexicalised
parser and annotation projection with crowdsourced
parallel data. Within both approaches we test two
parsing frameworks; the MaltParser
        <xref ref-type="bibr" rid="ref17 ref18">(Nivre et al.,
2007b)</xref>
        and the more recent UDPipe
        <xref ref-type="bibr" rid="ref22 ref32 ref40">(Straka and
Strakova´, 2017)</xref>
        . Both parsers are provided with
tokenised input.
4.1
      </p>
      <sec id="sec-4-1">
        <title>Model Transfer Approach</title>
        <p>The delexicalised model transfer approach is
straightforward, working on the basis of POS tags only. For
the training, the words in the Standard German corpus
are replaced by their POS tags. Accordingly, at
parsing time the Swiss German words are replaced by their
POS tag before parsing and re-inserted afterwards.
4.1.1</p>
      </sec>
      <sec id="sec-4-2">
        <title>POS tagging</title>
        <p>Part-of-speech tagging is an important step prior to
parsing because the syntactic structure builds upon the
5The Levenshtein distance (Levenshtein, 1966) measures the
difference between two sequences of characters. Hence, the
minimal edit distance between two words is the minimum number of
characters to be changed (i.e. inserted, deleted or substituted), in
order to make them equal.
POS information. Obviously, when training
delexicalised parsers, this step is crucial as the tags are the
only information available to the parser.</p>
        <p>
          For POS tagging Swiss German sentences, we used
the Wapiti
          <xref ref-type="bibr" rid="ref13">(Lavergne et al., 2010)</xref>
          model trained on
Release 2.2 of the NOAH corpus, where average
accuracy in 10-fold crossvalidation is 92.25%.
        </p>
        <p>The CoNLL-format includes UPOS tags in addition
to the fine-grained language-specific POS tags S(TTS
in the case of German and Swiss German). We used
the mapping provided by the UD project in order to
infer the UPOS tags from the given STTS tags.
4.2</p>
      </sec>
      <sec id="sec-4-3">
        <title>Annotation Projection Approach</title>
        <p>
          Annotation projection is not only more complex in
processing compared to model transfer but also needs
more resources. Most importantly, annotation
projection requires a word-aligned parallel corpus. Starting
from the crowdsourced sentences which are
sentencealigned, it is the task of a word aligner to compute
the most probable word alignments, i.e. the
information about which word of the (Swiss German) source
sentence corresponds to which word of the target
sentence, i.e. the translation. There are many tools for
this as it is a basic step also in machine translation
systems. We tested three of them: GIZA++
          <xref ref-type="bibr" rid="ref19">(Och and
Ney, 2003)</xref>
          , FastAlign
          <xref ref-type="bibr" rid="ref9">(Dyer et al., 2013)</xref>
          and
Monolingual Greedy Aligner (MGA)
          <xref ref-type="bibr" rid="ref21">(Rosa et al., 2012)</xref>
          .
        </p>
        <p>The idea of the annotation projection process is to
use the tool (here: parser) of a resource-rich language
on that language (here: German) and then project the
generated information (here: universal dependency
structures) along the word alignment to the target
language (here: Swiss German). In practice, this means
we train the parser on the Standard German
treebank (see Section 3.1) and parse the Standard German
translations of the Swiss German original sentences.
Then we project the resulting parse structure along the
word alignments from the German word to the
corresponding Swiss German word.
4.2.1</p>
      </sec>
      <sec id="sec-4-4">
        <title>Transfer of the Annotation</title>
        <p>The transfer is the core component of annotation
projection. The parse of the Standard German
translation is projected along the word alignment to its
Swiss German correspondent. The input consists of
the Standard German parse and the alignment between
the Standard German sentence and its Swiss German
version (GSW:DE). Algorithm 1 describes the
projection process.</p>
        <sec id="sec-4-4-1">
          <title>Data: DE parse &amp; alignment GSW:DE</title>
          <p>Result: DE parse transferred to GSW
for word alignment in sentence do
if 1:1 alignment then</p>
          <p>transfer parse of DE
else if 1:0 alignment (i.e. no DE word
aligned) then
attach GSW word to root as POS tag</p>
          <p>ADV and dependency label advmod
else 1:n alignment (i.e. several DE words
aligned)
transfer parse of aligned DE word with
smallest edit (Levenshtein) distance
end
end</p>
        </sec>
        <sec id="sec-4-4-2">
          <title>Algorithm 1: Transfer of parses.</title>
          <p>The case of 1:1 alignment where exactly one
German word is aligned to the Swiss German word is
easy; the only thing to do is projecting the
dependency of the German word to the Swiss German word.
If, however, there are several German words aligned
to one Swiss German word (1:n), the algorithm has
to decide which parse to transfer.6 In order to take
this decision, the algorithm computes the Levenshtein
6Note that the case of a 1:n alignment between a GSW word
and a DE multiword expression is not covered in this approach.
distances (Levenshtein, 1966) between the Swiss
German word and every aligned German word and takes
the one with the smallest edit distance. The most
challenging case is when no German word is aligned to
the Swiss German token. A simple baseline approach
attaches the corresponding Swiss German word as
adverbial modifier to the root of the sentence.</p>
          <p>The decision to treat every unaligned Swiss
German word as an adverb is taken on the basis of the
frequency distribution of POS tags; ADV is the second
most frequent POS tag (after NN) in the Swiss German
data. However, taking into consideration the word
itself, some more sophisticated rules can be
elaborated. Considering the differences between Standard
German and Swiss German as described by
Hollenstein and Aepli (2014), we can expect some words
like infinitive particles (PTKINF) (e.g. go) or the past
participle gsi (been) to remain unaligned. The
former because these words do not exist in Standard
German, the latter because Standard German simple past
tense is expressed by perfect tense in Swiss German,
typically resulting in a “spare” past participle in the
alignment. Furthermore, there are unaligned articles
because Swiss German requires articles in front of
proper names. Also punctuation including the
apostrophe is a source of errors which can easily be
corrected. The application of these more elaborate rules
have an impact of around 2 points on the evaluation
scores.</p>
          <p>Algorithm 1 transfers the German parses as they
are, as a consequence the numbering of the token IDs
is mixed up. Correcting the token IDs to be in
ascending order (from 1 to the length of the sentence)
requires the corresponding adjustment of the head
references. Furthermore, one needs to make sure that
there is exactly one root in a sentence.</p>
          <p>end</p>
        </sec>
        <sec id="sec-4-4-3">
          <title>Data: transferred DE parse to GSW words Result: valid GSW parse</title>
          <p>for sentence in parse do
if DE root was not projected to GSW parse
then</p>
          <p>take 1st VERB as root, else 1st NOUN
else if head of a projected word was not
projected to GSW parse then</p>
          <p>attach it to the root
end
Algorithm 2: Correction of transferred parses.</p>
          <p>Algorithm 2 goes through every sentence of the
input file and first makes sure that there is one root for
the sentence. If the root of the Standard German parse
has not been transferred to the Swiss German
sentence (missing word alignment), the first verb (UPOS
VERB) is taken as root and if there is no VERB in the
sentence, the firstNOUN is considered the root.
4.3</p>
        </sec>
      </sec>
      <sec id="sec-4-5">
        <title>Optimisation</title>
        <p>We tested two approaches for optimisation;
preprocessing of the training set and postprocessing rules to
be applied after parsing.
4.3.1</p>
      </sec>
      <sec id="sec-4-6">
        <title>Preprocessing</title>
        <p>One frequent mistake mostly observed in the
delexicalised approach is the wrong assignment of passive
dependency labels instead of their active counterpart.
The passive construction in Standard German is built
with the auxiliary werden, which can, however, also
be used in non-passive constructions. The
combination of VA* and a perfect participle (VVPP) is very
frequent in Swiss German, however, it is usually not a
passive construction but rather a perfect tense.
Therefore, a simple but effective solution is the introduction
of a new “set” of POS tags in the German UD training
set: VWFIN, VWINF, VWPP for finite verbs, infinitives
and participles respectively of the verb werden. This
means, all occurrences of the lemma werden as an
auxiliary (i.e. UPOS: AUX and STTS: VA{INF|PP})
are replaced by VW{INF|PP}. In this way, the system
learns to discriminate between the usage of werden as
auxiliary versus the usage as full verb and, most of all,
it learned to differentiate between the auxiliary
werden and the other auxiliaries haben (to have) and sein
(to be). Hence, the number of wrongly assigned
passive dependency labels decreased, which leads to an
improvement of around 2.5 to 3.5 points as presented
in Section 5.
4.3.2</p>
      </sec>
      <sec id="sec-4-7">
        <title>Postprocessing</title>
        <p>Some of the errors can easily be corrected with
simple rules in a postprocessing step. One example is a
frequent error caused by a remnant of the 1st UD
version which is handled differently in UD version 2. The
two labels oblique nominal (obl) and nominal
modifier (nmod) are confused because the latter was used
to modify nominals and predicates in UD v1.
However, in UD v2, obl is used for a nominal functioning
as an oblique argument, while nmod is used for
nominal dependents of another noun (phrase) only. This
means, if the head is a verb, adjective or adverb, the
dependency label has to be obl. If, instead, the head
is a noun, pronoun, name or number, the dependency
label is nmod.
5</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Results &amp; Discussion</title>
      <p>This section presents the different settings and
combinations of aforementioned resources, approaches and
tools. For the evaluation, we manually created a gold
standard consisting of 100 Swiss German sentences
taken from the resources presented in Section 3.2.
We evaluated the approaches according to Labelled
Attachment Score (LAS) and Unlabelled Attachment
Score (UAS)7, not excluding punctuation. The results
we present here are macro accuracy scores, that is,
the scores are computed separately for each sentence
and then averaged8. Note that there is a mismatch
in the actual annotation of punctuation between the
the Standard German UD treebank v2 and the official
guidelines we were applying. This difference in the
punctuation dependencies has an effect on the scores,
i.e. it lowers the scores presented here. Furthermore,
note that the test set containing 100 gold standard
sentences is small and therefore these results have to be
taken with a grain of salt.
5.1</p>
      <sec id="sec-5-1">
        <title>German Parser Accuracy</title>
        <p>In order to put the results into context, we checked
the performance of the parsers on the German UD v2
treebank using their split of training and test set. In
this setting, we left all the available information for
the parser to use, including morphology and lemmas.
The APPRART splitting is undone for the CoNLL-X
MaltParser input, not so for the UDPipe which takes
CoNLL-U as input format (and performs worse with
the MaltParser-CoNLL-X input). MaltParser reaches
a LAS of 79.71%, UDPipe 70.31% respectively.
5.2</p>
      </sec>
      <sec id="sec-5-2">
        <title>Direct Cross-lingual Parsing</title>
        <p>As a comparison to the main approaches, we applied
Standard German parsers directly to Swiss German.</p>
        <p>7UAS is the percentage of tokens with the correct syntactic
head, LAS the percentage of tokens assigned the correct syntactic
head as well as the correct dependency label.</p>
        <p>8Macro accuracy scores as opposed to the word-based micro
scores, where the true positives are summed up over the whole
treebank and divided by the total number of words.</p>
        <p>
          This means, we used the training set of the German
UD treebank to train the MaltParser (using
MaltOptimizer to get the best hyperparameter settings) and
UDPipe. Before training, we removed the
morphology and lemma information because this information
is not available in the Swiss German test set and
therefore the parsers cannot rely on it. Furthermore, for
the MaltParser we converted the training set from
CoNLL-U to CoNLL-X format because
MaltOptimizer cannot handle the former. Testing the
MaltParser model on the gold standard with
automatically assigned POS tags by Wapiti results in an LAS
of 55.28%. UDPipe only reaches 21.19% LAS, one
reason for this low accuracy could be that UDPipe
relies on word embedding information
          <xref ref-type="bibr" rid="ref22 ref32 ref40">(Straka and
Strakova´, 2017)</xref>
          , which results in a low recall when
applying a model trained on German to Swiss
German.
5.3
        </p>
      </sec>
      <sec id="sec-5-3">
        <title>Delexicalised Model Transfer</title>
        <p>Instead of giving the parser the Standard German
words as input like in the direct cross-lingual
approach, in the delexicalised approach we provide the
parser with POS information only. This means, the
words are replaced by STTS POS tags while all the
other columns stay the same. Given the small
evaluation set and a negligible difference in the results,
the two parsers’ performance can be considered the
same: ∼57% LAS for both when trained on the
preprocessed training set, i.e. differentiating the auxiliary
werden vs. the auxiliaries haben (to have) and sein (to
be) (see Section 4.3.1).
5.4</p>
      </sec>
      <sec id="sec-5-4">
        <title>Annotation Projection</title>
        <p>The results for the annotation projection approach
vary substantially depending on the combination of
aligner and parser. Starting from 46.45% LAS
(MaltParser + Fastalign), the combination of UDPipe and
Monolingual Greedy Aligner scores best in this
approach with 53.39% LAS. This score is reached with
the baseline transfer rules where unaligned words are
simply attached to the root as adverbs. Applying more
elaborate transfer rules (Section 4.2.1) results in an
improvement of 2.09 points to 55.65% LAS. The
preprocessing step does not improve the results in this
approach. These results show that the Monolingual
Greedy Aligner performs best in the task of DE/GSW
alignment. MGA takes character-based word
similarity into account which intuitively makes sense as the
information about similar letters is valuable
information when dealing with closely related languages such
as Standard German and Swiss German.
5.5</p>
      </sec>
      <sec id="sec-5-5">
        <title>Postprocessing</title>
        <p>The postprocessing rules do not show a huge impact
on the parsing results; the nmod/obl confusions for
example are still present. The reason for this is that
the parser assigned wrong heads to many of the words
and therefore the rule to correct the nmod/obl
confusions does not work. The LAS scores improve by
1.62 points for the cross-lingual MaltParser and 2.07
points for delexicalised model transfer and annotation
projection UDPipe approachs respectively, reaching
nearly 60% LAS accuracy.
5.6</p>
      </sec>
      <sec id="sec-5-6">
        <title>Discussion</title>
        <p>Table 1 shows the best results including the
corresponding setting for every approach. The best LAS
results of all the applied approaches are very close,
hence there is no clear answer to the question of which
approach works best. Annotation projection is the
most laborious among the three and as such not the
first option to choose. Furthermore, the transfer of the
annotation is strongly dependent on the performance
of the aligner, which in turn benefits from big parallel
corpora to be trained on. However, such big parallel
corpora do not exist yet for Swiss German dialects.</p>
        <p>Contrary to our expectations, training specific
models for different dialects does not have a huge impact
on the results. The word ordering for the St Gallen
dialect is closer to the Standard German word ordering
while Bernese dialect speakers often change the order
of the verbs. Due to these differences, we expected the
model transfer approach to perform worse on the Bern
dialect than annotation projection, where the word
order changes should be handled by the aligner.
Looking specifically at Bernese sentences with “switched”
word order (e.g. ha aafo gra¨nne (’I started to cry’),
gfunge hei gha (’have found’), het u¨bercho (’have
gotten’)), there is no significant difference between the
two approaches in our test set.
5.6.1</p>
      </sec>
      <sec id="sec-5-7">
        <title>Swiss German Variability</title>
        <p>The results presented here are not perfect and
certainly require further improvement in order for a
system to be used in real-life applications. Compared
with the Standard German parser accuracy, which
reaches almost 80% LAS on the German UD v2 with
standard settings of the parsers, there is room for
improvement. However, these numbers have to be set in
relation to the data we worked with. Even though we
could make use of Swiss German novels and
crowdsourced data, it is still a small data set. Furthermore,
the enormous spelling variability in Swiss German
dialects poses a serious challenge for all tools.
Statistical tools work best if the observed events are frequent.
However, they do not work well with sparse data
consisting of a large amount of hapax legomena, i.e. word
form which appears only once. Figure 4 shows the
frequencies (on y-axis) of type frequencies (x) in a Swiss
German text collection9 consisting of 6,155 sentences
with 105,692 tokens and 20,882 unique token types.
14,099 types appear only once (i.e. hapax legomena),
2,804 appear twice (i.e. hapax dislegomena) 19,874
less than 10 times and 20,767 less than 100 times.
5.7</p>
      </sec>
      <sec id="sec-5-8">
        <title>Silver Treebank Parsing Model</title>
        <p>Following the direct cross-lingual parsing
approach, we automatically parse 6,155 Swiss German
sentences9 in order to create a silver treebank. A
silver standard treebank, as opposed to a gold standard
treebank which is assumed to be correctly annotated,
is automatically annotated and may therefore contain
errors. Then, we use this silver treebank to train a
monolingual Swiss German parser and hence, create
a first monolingual Swiss German dependency
parsing model. The advantage of using a silver treebank
is the fact that it becomes a monolingual task.
However, this comes with the price of a faulty training set,
which is not the best resource to build a parser.</p>
        <p>9NOAH corpus plus 396 sentences from novels by Pedro Lenz
and Renato Kaiser, excluding gold standard sentences.</p>
        <p>Interestingly, the performance of the MaltParser
trained on the silver treebank reaches the same
performance as the direct cross-lingual parsing approach
itself, which was used to generate the silver treebank:
LAS 57.10%. Given that 6,000 sentences do not
constitute a large training set for a statistical parser, a
parser could probably profit from additional related
Standard German material. However, combining the
two training sets, i.e. the German Universal
Dependency treebank and the silver treebank gives slightly
worse results (LAS 55.46%).
There are several opportunities for further
improvement. Concerning the annotation projection approach,
the crucial alignment information needs to be
improved for example by ensembling over results from
different word aligners. In cases where alignment
does not work, adding further transfer and
postprocessing rules would be important. In addition, a
spelling normalisation strategy can help to deal with
the data sparseness imposed by the phonetic and
orthographic variability in Swiss German dialects.
Moreover, the outputs of the three parsing approaches
could be ensembled, e.g. via majority vote like for
alignment as aforementioned, to get rid of the
weaknesses of each approach. Furthermore, the silver
treebank created could be manually corrected in order to
generate a treebank which can be used as training set
for a monolingual dependency parser for Swiss
German. Finally, once the data sparseness for Swiss
German varieties is mitigated, modern neural methods
are promising as shown for example in the work by
Ammar et al. (2016). Ammar et al. train one
multilingual model that can be used to parse sentences
in several languages. In order to do so, they use
many resources including a bilingual dictionary for
adding cross-lingual lexical information, and a
monolingual corpus for training word embeddings. Such
approaches need a big amount of data of the language
to be parsed, which are still not available for Swiss
German.
6</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>Conclusion</title>
      <p>In this work, we experimented with a variety of
crosslingual approaches for parsing texts written in Swiss
German. For statistically driven systems, languages
with non-standardised orthography are a demanding
task. Swiss German dialects feature challenging
Natural Language Processing (NLP) problems with their
lack of orthographic spelling rules and a huge
pronunciation variety. This is a situation which leads to a
high degree of data sparseness and with it, a lack of
resources and tools for NLP.</p>
      <p>We tested a lexicalised annotation projection
method as well as a delexicalised model transfer
method. The annotation projection method requires
parallel sentences in both the resource-rich and the
low-resourced language while the delexicalised model
transfer approach only requires a monolingual
treebank of a closely related resource-rich language.</p>
      <p>The evaluation on a manually annotated gold
standard consisting of 100 sentences shows a 60%
Labelled Attachment Score (LAS) with negligible
differences between the different parsing approaches.
However, the annotation projection approach is more
complex than model transfer due to the transfer rules and
the crucial word alignment process.</p>
      <p>This work provides a first substantial step towards
closing a big gap in Natural Language Processing
tools for Swiss German and provides data10 to work
on further improvements.</p>
    </sec>
    <sec id="sec-7">
      <title>Acknowledgments</title>
      <p>We thank the AGORA project “Citizen Linguistics”
for making their translation data available and in
particular all our volunteer translators. We also thank
Renato Kaiser and Pedro Lenz for their permission to use
their novels in our experiments.</p>
      <p>10https://github.com/noe-eva/SwissGermanUD
forschung/ressourcen/lexika/TagSets/
stts-1999.pdf.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          Noe¨mi Aepli, Tanja Samardz˘ic´,
          <source>and Ruprecht von Waldenfels</source>
          .
          <year>2014</year>
          .
          <article-title>Part-of-Speech Tag Disambiguation by Cross-Linguistic Majority Vote</article-title>
          . In First Workshop on Applying NLP Tools to Similar Languages,
          <article-title>Varieties and Dialects (VarDial)</article-title>
          . Dublin.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          <string-name>
            <given-names>Waleed</given-names>
            <surname>Ammar</surname>
          </string-name>
          , George Mulcaire, Miguel Ballesteros, Chris Dyer, and
          <string-name>
            <surname>Noah</surname>
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Smith</surname>
          </string-name>
          .
          <year>2016</year>
          .
          <article-title>Many languages, one parser</article-title>
          .
          <source>TACL</source>
          <volume>4</volume>
          :
          <fpage>431</fpage>
          -
          <lpage>444</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          <string-name>
            <given-names>Reto</given-names>
            <surname>Baumgartner</surname>
          </string-name>
          .
          <year>2016</year>
          .
          <article-title>Morphological analysis and lemmatization for Swiss German using weighted transducers</article-title>
          . In Stefanie Dipper, Friedrich Neubarth, and Heike Zinsmeister, editors,
          <source>Proceedings of the 13th Conference on Natural Language Processing (KONVENS</source>
          <year>2016</year>
          ). Bochum, Germany.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          <string-name>
            <given-names>Sabine</given-names>
            <surname>Buchholz</surname>
          </string-name>
          and
          <string-name>
            <given-names>Erwin</given-names>
            <surname>Marsi</surname>
          </string-name>
          .
          <year>2006</year>
          .
          <article-title>Conll-x shared task on multilingual dependency parsing</article-title>
          .
          <source>In In Proceedings of CoNLL</source>
          . pages
          <fpage>149</fpage>
          -
          <lpage>164</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          <string-name>
            <surname>Marie-Catherine de Marneffe</surname>
            , Timothy Dozat, Natalia Silveira, Katri Haverinen, Filip Ginter, Joakim Nivre, and
            <given-names>Christopher D.</given-names>
          </string-name>
          <string-name>
            <surname>Manning</surname>
          </string-name>
          .
          <year>2014</year>
          .
          <article-title>Universal stanford dependencies: A cross-linguistic typology</article-title>
          .
          <source>In Proceedings of the Ninth International Conference on Language Resources</source>
          and
          <article-title>Evaluation (LREC-</article-title>
          <year>2014</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          <string-name>
            <surname>Marie-Catherine de Marneffe</surname>
          </string-name>
          , Bill MacCartney, and
          <string-name>
            <surname>Christopher</surname>
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Manning</surname>
          </string-name>
          .
          <year>2006</year>
          .
          <article-title>Generating typed dependency parses from phrase structure parses</article-title>
          .
          <source>In 5th International Conference on Language Resources and Evaluation (LREC</source>
          <year>2006</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          <string-name>
            <surname>Marie-Catherine de Marneffe and Christopher D. Manning</surname>
          </string-name>
          .
          <year>2008</year>
          .
          <article-title>The Stanford typed dependencies representation</article-title>
          .
          <source>In COLING Workshop on Cross-framework and Crossdomain Parser Evaluation.</source>
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          <source>Christa Du¨rscheid and Elisabeth Stark</source>
          .
          <year>2011</year>
          .
          <article-title>SMS4science: An international corpus-based texting project and the specific challenges for multilingual Switzerland</article-title>
          . Digital Discourse: Language in the New Media pages
          <fpage>299</fpage>
          -
          <lpage>320</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          <string-name>
            <given-names>Chris</given-names>
            <surname>Dyer</surname>
          </string-name>
          , Victor Chahuneau, and
          <string-name>
            <surname>Noah</surname>
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Smith</surname>
          </string-name>
          .
          <year>2013</year>
          .
          <article-title>A simple, fast, and effective reparameterization of IBM model 2</article-title>
          .
          <string-name>
            <surname>In</surname>
            <given-names>HLT</given-names>
          </string-name>
          -NAACL. pages
          <fpage>644</fpage>
          -
          <lpage>648</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          <string-name>
            <given-names>Nora</given-names>
            <surname>Hollenstein</surname>
          </string-name>
          and Noe¨mi Aepli.
          <year>2014</year>
          .
          <article-title>Compilation of a swiss german dialect corpus and its application to pos tagging</article-title>
          .
          <source>In Proceedings of the First Workshop on Applying NLP Tools to Similar Languages, Varieties and Dialects</source>
          . Dublin, Ireland, pages
          <fpage>85</fpage>
          -
          <lpage>94</lpage>
          . http://www.aclweb.org/anthology/W14-5310.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          <source>German Society for Computational Linguistics and Language Technology. German Society for Computational Linguistics and Language Technology</source>
          , Duisburg-Essen, Germany, pages
          <fpage>108</fpage>
          -
          <lpage>109</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          <string-name>
            <given-names>Rebecca</given-names>
            <surname>Hwa</surname>
          </string-name>
          , Philip Resnik, Amy Weinberg, Clara Cabezas, and
          <string-name>
            <given-names>Okan</given-names>
            <surname>Kolak</surname>
          </string-name>
          .
          <year>2005</year>
          .
          <article-title>Bootstrapping parsers via syntactic projection across parallel texts</article-title>
          .
          <source>Natural language engineering</source>
          <volume>11</volume>
          (03):
          <fpage>311</fpage>
          -
          <lpage>325</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          <string-name>
            <given-names>Thomas</given-names>
            <surname>Lavergne</surname>
          </string-name>
          , Olivier Cappe´, and Franc¸ois Yvon.
          <year>2010</year>
          .
          <article-title>Practical Very Large Scale CRFs</article-title>
          .
          <source>In Proceedings the 48th Annual Meeting of the Association for Computational Linguistics (ACL)</source>
          .
          <source>Uppsala, Sweden</source>
          , pages
          <fpage>504</fpage>
          -
          <lpage>513</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          <string-name>
            <surname>Ryan</surname>
            <given-names>McDonald</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>Joakim</given-names>
            <surname>Nivre</surname>
          </string-name>
          ,
          <string-name>
            <surname>Yvonne</surname>
            <given-names>QuirmbachBrundage</given-names>
          </string-name>
          , Yoav Goldberg,
          <string-name>
            <surname>Dipanjan Das</surname>
          </string-name>
          ,
          <string-name>
            <surname>Kuzman Ganchev</surname>
            , Keith Hall, Slav Petrov, Hao Zhang, Oscar Ta¨ckstro¨m, Claudia Bedini, Nu´ria Bertomeu Castello´, and
            <given-names>Jungmee</given-names>
          </string-name>
          <string-name>
            <surname>Lee</surname>
          </string-name>
          .
          <year>2013</year>
          .
          <article-title>Universal dependency annotation for multilingual parsing</article-title>
          .
          <source>In Proceedings of the 51st Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers)</source>
          . pages
          <fpage>92</fpage>
          -
          <lpage>97</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          <string-name>
            <given-names>Tahira</given-names>
            <surname>Naseem</surname>
          </string-name>
          , Harr Chen, Regina Barzilay, and
          <string-name>
            <given-names>Mark</given-names>
            <surname>Johnson</surname>
          </string-name>
          .
          <year>2010</year>
          .
          <article-title>Using universal linguistic knowledge to guide grammar induction</article-title>
          .
          <source>In Proceedings of the 2010 Conference on Empirical Methods in Natural Language Processing</source>
          . pages
          <fpage>1234</fpage>
          -
          <lpage>1244</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          <string-name>
            <given-names>Joakim</given-names>
            <surname>Nivre</surname>
          </string-name>
          ,
          <string-name>
            <surname>Marie-Catherine de Marneffe</surname>
            , Filip Ginter, Yoav Goldberg, Jan Hajic,
            <given-names>Christopher D.</given-names>
          </string-name>
          <string-name>
            <surname>Manning</surname>
          </string-name>
          ,
          <string-name>
            <surname>Ryan</surname>
            <given-names>McDonald</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>Slav</given-names>
            <surname>Petrov</surname>
          </string-name>
          , Sampo Pyysalo, Natalia Silveira, Reut Tsarfaty, and
          <string-name>
            <given-names>Daniel</given-names>
            <surname>Zeman</surname>
          </string-name>
          .
          <year>2016</year>
          .
          <article-title>Universal dependencies v1: A multilingual treebank collection</article-title>
          .
          <source>In Proceedings of the Tenth International Conference on Language Resources and Evaluation (LREC</source>
          <year>2016</year>
          ). Paris, France.
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          <string-name>
            <given-names>Joakim</given-names>
            <surname>Nivre</surname>
          </string-name>
          , Johan Hall, Sandra Ku¨bler,
          <string-name>
            <surname>Ryan</surname>
            <given-names>McDonald</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>Jens</given-names>
            <surname>Nilsson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Sebastian</given-names>
            <surname>Riedel</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Deniz</given-names>
            <surname>Yuret</surname>
          </string-name>
          .
          <year>2007a</year>
          .
          <article-title>The CoNLL 2007 shared task on dependency parsing</article-title>
          .
          <source>In Proceedings of the CoNLL Shared Task Session of EMNLP-CoNLL</source>
          <year>2007</year>
          . Prague, Czech Republic, pages
          <fpage>915</fpage>
          -
          <lpage>932</lpage>
          . http://www.aclweb.org/anthology/D/D07/D07-1096.
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          <string-name>
            <given-names>Joakim</given-names>
            <surname>Nivre</surname>
          </string-name>
          , Johan Hall, Jens Nilsson, Atanas Chanev, Gu¨l S¸en Eryigit,
          <article-title>Sandra K u¨bler, Svetoslav Marinov, and Erwin Marsi</article-title>
          . 2007b.
          <article-title>Maltparser: A languageindependent system for data-driven dependency parsing</article-title>
          .
          <source>Natural Language Engineering</source>
          <volume>13</volume>
          (
          <issue>2</issue>
          ):
          <fpage>95</fpage>
          -
          <lpage>135</lpage>
          . https://doi.org/10.1017/S1351324906004505.
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          <string-name>
            <given-names>Nora</given-names>
            <surname>Hollenstein</surname>
          </string-name>
          and Noe¨mi Aepli.
          <year>2015</year>
          .
          <article-title>A resource for natural language processing of swiss german dialects</article-title>
          .
          <source>In Proceedings of the International Conference of the Franz Josef Och and Hermann Ney</source>
          .
          <year>2003</year>
          .
          <article-title>A systematic comparison of various statistical alignment models</article-title>
          .
          <source>Computational Linguistics</source>
          <volume>29</volume>
          (
          <issue>1</issue>
          ):
          <fpage>19</fpage>
          -
          <lpage>51</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          <string-name>
            <given-names>Slav</given-names>
            <surname>Petrov</surname>
          </string-name>
          ,
          <string-name>
            <surname>Dipanjan Das</surname>
          </string-name>
          , and
          <string-name>
            <surname>Ryan McDonald</surname>
          </string-name>
          .
          <year>2012</year>
          .
          <article-title>A universal part-of-speech tagset</article-title>
          .
          <source>In Proceedings of the Eighth International Conference on Language Resources and Evaluation (LREC-2012)</source>
          . pages
          <fpage>2089</fpage>
          -
          <lpage>2096</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          <string-name>
            <given-names>Rudolf</given-names>
            <surname>Rosa</surname>
          </string-name>
          , Ondˇrej Dusˇek, David Marecˇek, and
          <string-name>
            <given-names>Martin</given-names>
            <surname>Popel</surname>
          </string-name>
          .
          <year>2012</year>
          .
          <article-title>Using parallel features in parsing of machine-translated sentences for correction of grammatical errors</article-title>
          .
          <source>In Proceedings of the Sixth Workshop on Syntax, Semantics and Structure in Statistical Translation. Jeju, Republic of Korea</source>
          , pages
          <fpage>39</fpage>
          -
          <lpage>48</lpage>
          . http://www.aclweb.org/anthology/W12-4205.
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          <string-name>
            <surname>Rudolf</surname>
            <given-names>Rosa</given-names>
          </string-name>
          , Daniel Zeman, David Marecˇek, and Zdenek Zˇabokrtsky´.
          <year>2017</year>
          .
          <article-title>Slavic forest, Norwegian wood</article-title>
          .
          <source>In Proceedings of the Fourth Workshop on NLP for Similar Languages, Varieties and Dialects (VarDial)</source>
          . Valencia, Spain, pages
          <fpage>210</fpage>
          -
          <lpage>219</lpage>
          . http://www.aclweb.org/anthology/W17-1226.
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          <string-name>
            <surname>Tanja</surname>
            <given-names>Samardzˇic´</given-names>
          </string-name>
          , Yves Scherrer, and
          <string-name>
            <given-names>Elvira</given-names>
            <surname>Glaser</surname>
          </string-name>
          .
          <year>2015</year>
          .
          <article-title>Normalising orthographic and dialectal variants for the automatic processing of Swiss German</article-title>
          . In Language and Technology Conference:
          <article-title>Human Language Technologies as a Challenge for Computer Science</article-title>
          and Linguistics. Poznan, Poland, pages
          <fpage>294</fpage>
          -
          <lpage>298</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          <string-name>
            <given-names>Tanja</given-names>
            <surname>Samardzic</surname>
          </string-name>
          , Yves Scherrer, and
          <string-name>
            <given-names>Elvira</given-names>
            <surname>Glaser</surname>
          </string-name>
          .
          <year>2016</year>
          .
          <article-title>ArchiMob - a corpus of spoken Swiss German</article-title>
          .
          <source>In Proceedings of the Tenth International Conference on Language Resources and Evaluation (LREC</source>
          <year>2016</year>
          ). Paris, France.
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          <string-name>
            <given-names>Yves</given-names>
            <surname>Scherrer</surname>
          </string-name>
          .
          <year>2007</year>
          .
          <article-title>Adaptive string distance measures for bilingual dialect lexicon induction</article-title>
          .
          <source>In Proceedings of the ACL 2007 Student Research Workshop</source>
          . Prague, Czech Republic, pages
          <fpage>55</fpage>
          -
          <lpage>60</lpage>
          . http://www.aclweb.org/anthology/P/P07/P07-3010.
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          <string-name>
            <given-names>Yves</given-names>
            <surname>Scherrer</surname>
          </string-name>
          .
          <year>2012</year>
          .
          <article-title>Machine translation into multiple dialects: The example of Swiss German</article-title>
          .
          <source>In 7th SIDG Congress - Dialect 2.0.</source>
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          <string-name>
            <given-names>Yves</given-names>
            <surname>Scherrer</surname>
          </string-name>
          .
          <year>2013</year>
          .
          <article-title>Continuous variation in computational morphology - the example of Swiss German</article-title>
          . In TheoreticAl and Computational MOrphology:
          <article-title>New Trends and Synergies (TACMO)</article-title>
          .
          <source>19th International Congress of Linguists</source>
          , Gene`ve, Suisse. http://hal.inria.fr/hal-00851251.
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          <string-name>
            <given-names>Yves</given-names>
            <surname>Scherrer</surname>
          </string-name>
          and
          <string-name>
            <given-names>Rambow</given-names>
            <surname>Owen</surname>
          </string-name>
          .
          <year>2010</year>
          .
          <article-title>Natural Language Processing for the Swiss German Dialect Area</article-title>
          .
          <source>In Proceedings of the Conference on Natural Language Processing (KONVENS)</source>
          .
          <source>Saarbru¨cken, Germany</source>
          , pages
          <fpage>93</fpage>
          -
          <lpage>102</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          <string-name>
            <given-names>Anne</given-names>
            <surname>Schiller</surname>
          </string-name>
          , Simone Teufel, Christine Sto¨ckert, and
          <string-name>
            <given-names>Christine</given-names>
            <surname>Thielen</surname>
          </string-name>
          .
          <year>1999</year>
          .
          <article-title>Guidelines fu¨r das Tagging deutscher Textkorpora mit STTS</article-title>
          . http://www.ims.uni-stuttgart.de/
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          <string-name>
            <given-names>Benjamin</given-names>
            <surname>Snyder</surname>
          </string-name>
          , Tahira Naseem, Jacob Eisenstein, and
          <string-name>
            <given-names>Regina</given-names>
            <surname>Barzilay</surname>
          </string-name>
          .
          <year>2008</year>
          .
          <article-title>Unsupervised multilingual learning for POS tagging</article-title>
          .
          <source>In Proceedings of the 2008 Conference on Empirical Methods in Natural Language Processing. Honolulu, Hawaii</source>
          , pages
          <fpage>1041</fpage>
          -
          <lpage>1050</lpage>
          . http://www.aclweb.org/anthology/D08-1109.
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          <string-name>
            <given-names>Elisabeth</given-names>
            <surname>Stark</surname>
          </string-name>
          , Simone Ueberwasser, and Anne Go¨hrig.
          <year>2014</year>
          .
          <article-title>Corpus ”What's up, Switzerland?”. www. whatsup-switzerland</article-title>
          .ch.
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          <string-name>
            <given-names>Milan</given-names>
            <surname>Straka</surname>
          </string-name>
          and Jana Strakova´.
          <year>2017</year>
          .
          <article-title>Tokenizing, pos tagging, lemmatizing and parsing ud 2.0 with udpipe</article-title>
          .
          <source>In Proceedings of the CoNLL</source>
          <year>2017</year>
          <article-title>Shared Task: Multilingual Parsing from Raw Text to Universal Dependencies</article-title>
          . Vancouver, Canada, pages
          <fpage>88</fpage>
          -
          <lpage>99</lpage>
          . http://www.aclweb.org/anthology/K/K17/K17- 3009.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref33">
        <mixed-citation>
          <string-name>
            <given-names>Oscar</given-names>
            <surname>Ta</surname>
          </string-name>
          <article-title>¨ckstro¨m, Ryan McDonald</article-title>
          ,
          <string-name>
            <given-names>and Joakim</given-names>
            <surname>Nivre</surname>
          </string-name>
          .
          <year>2013</year>
          .
          <article-title>Target language adaptation of discriminative transfer parsers</article-title>
          .
          <source>In Proceedings of the</source>
          <year>2013</year>
          <article-title>Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</article-title>
          . Atlanta, Georgia, pages
          <fpage>1061</fpage>
          -
          <lpage>1071</lpage>
          . http://www.aclweb.org/anthology/N13-1126.
        </mixed-citation>
      </ref>
      <ref id="ref34">
        <mixed-citation>
          <source>Jo¨rg Tiedemann</source>
          .
          <year>2014</year>
          .
          <article-title>Rediscovering annotation projection for cross-lingual parser induction</article-title>
          .
          <source>In Proceedings of COLING</source>
          <year>2014</year>
          ,
          <source>the 25th International Conference on Computational Linguistics: Technical Papers</source>
          . pages
          <fpage>1854</fpage>
          -
          <lpage>1864</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref35">
        <mixed-citation>
          <source>Jo¨rg Tiedemann</source>
          .
          <year>2015</year>
          .
          <article-title>Cross-lingual dependency parsing with universal dependencies and predicted PoS labels</article-title>
          .
          <source>In Proceedings of the Third International Conference on Dependency Linguistics (Depling</source>
          <year>2015</year>
          ). pages
          <fpage>340</fpage>
          -
          <lpage>349</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref36">
        <mixed-citation>
          <article-title>Jo¨rg Tiedemann, Zˇeljko Agic´</article-title>
          , and
          <string-name>
            <given-names>Joakim</given-names>
            <surname>Nivre</surname>
          </string-name>
          .
          <year>2014</year>
          .
          <article-title>Treebank translation for cross-lingual parser induction</article-title>
          .
          <source>In Proceedings of the Eighteenth Conference on Computational Natural Language Learning</source>
          . Ann Arbor, Michigan, pages
          <fpage>130</fpage>
          -
          <lpage>140</lpage>
          . http://www.aclweb.org/anthology/W14-1614.
        </mixed-citation>
      </ref>
      <ref id="ref37">
        <mixed-citation>
          <string-name>
            <given-names>David</given-names>
            <surname>Yarowsky</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Grace</given-names>
            <surname>Ngai</surname>
          </string-name>
          , and Richard Wicentowski.
          <year>2001</year>
          .
          <article-title>Inducing multilingual text analysis tools via robust projection across aligned corpora</article-title>
          .
          <source>In Proceedings of the first international conference on Human language technology research</source>
          . pages
          <fpage>1</fpage>
          -
          <lpage>8</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref38">
        <mixed-citation>
          <string-name>
            <given-names>Marcos</given-names>
            <surname>Zampieri</surname>
          </string-name>
          , Shervin Malmasi, Nikola Ljubesˇic´,
          <string-name>
            <surname>Preslav</surname>
            <given-names>Nakov</given-names>
          </string-name>
          , Ahmed Ali, Jo¨rg Tiedemann, Yves Scherrer, and Noe¨mi Aepli.
          <year>2017</year>
          .
          <article-title>Findings of the VarDial evaluation campaign 2017</article-title>
          .
          <source>In Proceedings of the Fourth Workshop on NLP for Similar Languages, Varieties and Dialects (VarDial)</source>
          . Valencia, Spain, pages
          <fpage>1</fpage>
          -
          <lpage>15</lpage>
          . http://www.aclweb.org/anthology/W17-1201.
        </mixed-citation>
      </ref>
      <ref id="ref39">
        <mixed-citation>
          <string-name>
            <given-names>D.</given-names>
            <surname>Zeman</surname>
          </string-name>
          and
          <string-name>
            <given-names>Philip</given-names>
            <surname>Resnik</surname>
          </string-name>
          .
          <year>2008</year>
          .
          <article-title>Cross-language parser adaptation between related languages</article-title>
          .
          <source>In Proceedings of the IJCNLP-08 Workshop on NLP for Less Privileged Languages</source>
          . pages
          <fpage>35</fpage>
          -
          <lpage>42</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref40">
        <mixed-citation>
          <string-name>
            <given-names>Daniel</given-names>
            <surname>Zeman</surname>
          </string-name>
          , Martin Popel, Milan Straka, Jan Hajic, Joakim Nivre, Filip Ginter, Juhani Luotolahti, Sampo Pyysalo, Slav Petrov, Martin Potthast, Francis Tyers, Elena Badmaeva, Memduh Gokirmak, Anna Nedoluzhko, Silvie Cinkova,
          <article-title>Jan Hajic jr</article-title>
          .,
          <string-name>
            <surname>Jaroslava</surname>
            <given-names>Hlavacova</given-names>
          </string-name>
          , Va´clava Kettnerova´,
          <string-name>
            <surname>Zdenka</surname>
            <given-names>Uresova</given-names>
          </string-name>
          , Jenna Kanerva, Stina Ojala, Anna Missila¨,
          <string-name>
            <surname>Christopher</surname>
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Manning</surname>
          </string-name>
          , Sebastian Schuster, Siva Reddy, Dima Taji, Nizar Habash, Herman Leung,
          <string-name>
            <surname>Marie-Catherine de Marneffe</surname>
          </string-name>
          , Manuela Sanguinetti, Maria Simi, Hiroshi Kanayama, Valeria dePaiva,
          <string-name>
            <surname>Kira</surname>
            <given-names>Droganova</given-names>
          </string-name>
          ,
          <article-title>He´ctor Mart´ınez Alonso, C¸a g˘r C¸o¨ltekin, Umut Sulubacak</article-title>
          , Hans Uszkoreit, Vivien Macketanz, Aljoscha Burchardt, Kim Harris, Katrin Marheinecke, Georg Rehm, Tolga Kayadelen, Mohammed Attia, Ali Elkahky, Zhuoran Yu, Emily Pitler, Saran Lertpradit, Michael Mandl, Jesse Kirchner, Hector Fernandez Alcalde, Jana Strnadova´,
          <string-name>
            <surname>Esha</surname>
            <given-names>Banerjee</given-names>
          </string-name>
          , Ruli Manurung, Antonio Stella, Atsuko Shimada, Sookyoung Kwak, Gustavo Mendonca, Tatiana Lando, Rattima Nitisaroj, and
          <string-name>
            <given-names>Josie</given-names>
            <surname>Li</surname>
          </string-name>
          .
          <year>2017</year>
          .
          <article-title>Conll 2017 shared task: Multilingual parsing from raw text to universal dependencies</article-title>
          .
          <source>In Proceedings of the CoNLL</source>
          <year>2017</year>
          <article-title>Shared Task: Multilingual Parsing from Raw Text to Universal Dependencies</article-title>
          . Vancouver, Canada, pages
          <fpage>1</fpage>
          -
          <lpage>19</lpage>
          . http://www.aclweb.org/anthology/K/K17/K17- 3001.pdf.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>