<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Exploring cross-lingual word embeddings for the inference of bilingual dictionaries?</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Universidade da Corun~a</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Grupo LyS</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Departamento de Letras Campus da Zapateira</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>A Corun~a</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Spain marcos.garcia.gonzalez@udc.gal</string-name>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Universidade da Corun~a, CITIC, Campus de Elvin~a</institution>
          ,
          <addr-line>15071 A Corun~a</addr-line>
          ,
          <country country="ES">Spain</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Universidade da Corun~a, Grupo LyS, Departamento de Computacion Campus de Elvin~a</institution>
          ,
          <addr-line>15071 A Corun~a</addr-line>
          ,
          <country country="ES">Spain</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>We describe four systems to generate automatically bilingual dictionaries based on existing ones: three transitive systems di ering only in the pivot language used, and a system based on a di erent approach which only needs monolingual corpora in both the source and target languages. All four methods make use of cross-lingual word embeddings trained on monolingual corpora, and then mapped into a shared vector space. Experimental results con rm that our strategy has a good coverage and recall, achieving a performance comparable to to the best submitted systems on the TIAD 2019 gold standard set among the teams participating at the TIAD shared task.</p>
      </abstract>
      <kwd-group>
        <kwd>Bilingual dictonaries cross-lingual word embeddings distributional semantics</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        Most research and development in natural language processing (NLP) and text
mining was initially conducted for English. In the last decades, there was a surge
in the application of NLP to other European languages as well as to the
mostspoken languages worldwide (e.g., Chinese [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ], Arabic [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]). Thus, in present
day, the tendency is not to focus on monolingual texts, but to try to develop
techniques and tools that can analyze texts written in di erent languages. In
this context, the availability of bilingual lexicons is a necessity, not only for
machine translation but for multilingual text mining tasks [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ], such as sentiment
analysis [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ] or information extraction and summarization [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. However, it is not
possible to get a manually constructed bilingual lexicon for any arbitrary pair
of languages. Hence the interest in de ning models and techniques to construct
bilingual language resources from those available for other pairs of languages.
      </p>
      <p>
        For this purpose, most methods rely on comparable corpora [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ], taking a
measure of common words in their contexts as rst clue to the similarity between
a word and its translation. However, in this type of resources the position and
the frequency of the source and target words are not comparable, and the
translation of a word might not exist in a given pair of comparable documents [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ].
Taking into account these limitations, several techniques have been tried, ranging
from adaptations of traditional Information Retrieval metrics [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] to unsupervised
multilingual word embeddings [
        <xref ref-type="bibr" rid="ref2 ref9">9, 2</xref>
        ]. In the case of rare words (terms that
appear from 1 to 5 times in a collection of documents), it is even possible to use
a vector representation of contexts and a classi er trained on one pair of
languages to extract translations for another pair of languages with a reasonable
performance [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ], but this technique does not scale to the full range of words.
      </p>
      <p>
        In [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ], Rapp proposes a methodology for extracting bilingual lexicons from
comparable corpora which is based on aligning comparable documents, using
multiple pivot languages and considering word senses rather than words to solve
ambiguities. Usually, taking into account word senses requires the existence of
a previously compiled dictionary with all the possible senses for each word, in
the style of WordNet, but since the construction of this kind of dictionaries is
expensive even when reusing existing resources [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], they will not be available
for most pairs of languages. An alternative consists in projecting words into
a high-dimensional concept space: each word is converted into a vector of real
numbers such that words having similar vectors are supposed to correspond to
similar meanings. This, in brief, is the foundation of our proposal for using word
embeddings on comparable corpora for the construction of bilingual dictionaries.
2
      </p>
    </sec>
    <sec id="sec-2">
      <title>Description of the LyS systems</title>
      <p>The objective of TIAD shared task is to generate automatically bilingual
dictionaries based on existing ones. In particular, participants were required to
generate new translations automatically between English (EN), French (FR)
and Portuguese (PT) based on known translations contained in the Apertium
RDF graph4. As these three languages are not directly connected in this graph,
no translations can be obtained directly among them. The use of other freely
available sources of background knowledge to improve performance is allowed,
as long as no direct translation among the target language pairs is applied.</p>
      <p>We presented four similar systems to the TIAD 2019 shared task: three
transitive systems di ering only in the pivot language used, and a system based on</p>
      <sec id="sec-2-1">
        <title>4 http://linguistic.linkeddata.es/apertium/</title>
        <p>
          a di erent approach which only needs monolingual corpora in both the source
and target languages. Every system makes use of cross-lingual word embeddings
trained on monolingual corpora, and then mapped into a shared vector space
using VecMap [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ]. Apart from the three source{target languages of the shared task
(EN, PT, and FR), we also evaluated two additional pivot languages: Spanish
(ES), and Catalan (CA).
2.1
        </p>
        <p>Data processing
In order to obtain similar models, and also to cover a general vocabulary in each
case, we decided to employ the di erent editions of Wikipedia as corpora for
learning the word embeddings. For English, Portuguese, and Spanish, we used
the Wikipedia dumps from January 2018, while for French and Catalan we have
downloaded the January 2019 data.</p>
        <p>
          Since we work with dictionary data, we have pre-processed each corpus to
obtain morphosyntactic information as well as to reduce the vocabulary size of
each model. Thus, we can obtain the PoS (Part of Speech) tag of each word
and avoid the extraction of in ected forms that do not appear in dictionaries.
This process was carried out using LinguaKit [
          <xref ref-type="bibr" rid="ref6 ref7">7, 6</xref>
          ] for Spanish, Portuguese, and
English, and FreeLing [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ] for Catalan and French. These processed data were
then converted into lemma PoSTag corpora to train distributional models with
this lexical and morphosyntactic information.
        </p>
        <p>
          After that, we used these modi ed corpora to train distributional semantics
models using word2vec [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ]. We took advantage of the gensim implementation
[
          <xref ref-type="bibr" rid="ref16">16</xref>
          ] to train skip-gram models with 300 dimensions using windows of 5 tokens
and a frequency threshold of 5. Table 1 contains the size of each Wikipedia
(in number of tokens) as well as the number of lemma PoSTag entries of the
word2vec vocabularies. It is worth mentioning that the large size of the English
vocabulary may reduce the computational e ciency of our approach when
combined with other large models such as the French one, so we decided to train an
additional English model (EN10 ) selecting only those lemma PoStag elements
with 10 or more occurrences in the corpus.
        </p>
        <p>
          Once we built the monolingual models for each language, we mapped them
into shared bilingual spaces using VecMap [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ], a framework to learn
crosslingual embeddings mappings. We applied the semi-supervised approach to each
language pair, thus obtaining bilingual word-embeddings models. For EN/FR,
FR/PT, and PT/EN, we only used 100 digits in lemma PoSTag format (e.g.,
0 NUM, 1 NUM, . . . , 99 NUM) as seed dictionary, but we did not take
advantage of any bilingual resource. For the other pairs, we used these 100 numbers as
well as 300 randomly selected words (100 adjectives, 100 nouns, and 100 verbs)
which were automatically translated and then reviewed by the authors.
2.2
        </p>
        <p>Algorithm
With a view to exploring solely the performance of cross-lingual word
embeddings in this task, it must be noted that our approach only uses the rst and
last columns of the input data (the source lemma and its PoS tag), so we do
not utilize information such as the sense of each entry or its translation in other
languages. Besides, the strategy was principally designed to obtain translations
of single lexical words, so multiword expressions (specially compound proper
nouns and non-compositional expressions such as idioms) are not well covered
by the algorithm. These are, however, interesting cases for further research.</p>
        <p>To translate between a source (src) and a target (tgt ) language, our algorithm
uses a pivot language (pvt ) relying on the referred cross-lingual word embeddings
and computing the similarity by means of the cosine distance between the two
word vectors. We take the following steps:
1. For each single word (lemma PoSTag, except proper nouns) in the input
dictionary (e.g., EN-GL), we rst check whether it appears in our source
vocabulary with the same PoS tag. In this case, we select the two most
similar entries with the same PoS tag in the pivot vocabulary. Then, for
each of these words we get the two closest entries in the target model (also
with the same PoS tag). If any of the words appears in the models with
the same morphosyntactic category, the most similar entry (adjective, noun,
verb, or adverb) is selected. At the end of this process, we have 0 to 4
candidate translations for each input word, together with a con dence value
(the cosine distance between the words in the pvt and tgt models). If no
translation is found, the algorithm applies a default rule which uses the
input lemma as the translation with a 0 con dence value.
2. For single proper nouns (composed by just one word), we rst check whether
they appear in the pivot language (both in upper-case and lower-case) with a
similarity greater than 0.5. In this case, we use the same procedure between
pvt and tgt, selecting the target entry with a con dence value of 1. If the
input proper noun does not appear in the pivot or target languages, we
simply select the most similar proper noun (from the closest 50 words). If
no proper noun is found, we apply the default rule.
3. To translate multiword expressions (MWEs) we apply a basic approach
which uses a list of MWEs in the target language extracted from the
input dictionaries (mwe-list ). Thus, for each input MWE, we select its two
longest words (src 1 and src 2 ), and applying a similar strategy to that of
single words, we select two candidate translations for each one: w1 and w1b
for src 1, and w2 and w2b for src 2. Then, using the mwe-list in the target
language, we select the rst MWE which contain both w1 and w2.
Otherwise, we look for expressions containing other combinations of the candidate
translations. If no translation is found, we also apply the default rule.</p>
        <p>The output of this process contains, for each input entry, 1 to 4 candidate
translations with their con dence value. In a post-processing step we select the
rst candidate as the target translation for lemma PoSTag pairs with only one
sense in each input dictionary. However, as our method entirely relies on the
input word and PoS tag (and not on its translation or its semantic information),
it produces the same output for homonyms and the di erent senses of polysemous
words. To alleviate this issue, the post-processing step respectively selects the
third, second, and fourth candidates (if any) as the translations of other entries
of the same word form.</p>
        <p>Using di erent pivot languages, we applied this algorithm in four runs:
{ LyS: this run uses, for each translation direction, the third language of the
shared task as pivot. Thus, for EN!FR, Portuguese was the pivot language,
while for FR!PT and PT!EN, English and French were the pivot
languages, respectively.
{ LyS-CA: in this system, Catalan was used as the pivot language for each
translation.
{ LyS-ES: this approach is identical to the previous one, but with Spanish
(instead of Catalan) as pivot.
{ LyS-DT: this last strategy does not use a pivot language, but only the
source and target cross-lingual word embeddings models. As mentioned, and
to follow the guidelines of the shared task, the mapping of these models was
carried out without any bilingual information.</p>
        <p>Except for the EN$FR translations (where we used the EN10 model), the
full English model was used in all the other directions (see Table 1).
3</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Results</title>
      <p>Evaluation of the results was carried out by the organisers against a
goldstandard set built by extracting translations from manually compiled pairs of
K Dictionaries5 (KD). As the coverage of KD is not the same as Apertium, the
subset covered by Apertium was taken to build the gold standard, i.e., those KD
translations for which the source and target terms are present in both Apertium
RDF source and target lexicons. The gold standard set was not available to
participants, so that we could not carry out a systematic error analysis which may
be useful for further research.</p>
      <p>Due to a misunderstanding, our team, like other participants, understood
that evaluation would be performed in the direction EN!FR, FR!PT and</p>
      <sec id="sec-3-1">
        <title>5 https://www.lexicala.com/resources#dictionaries</title>
        <p>0.7
0.6
0.5
0.4
0.3
0.2
0.1
0.00.0</p>
        <p>Coverage
Precision
F1
Recall</p>
        <p>0.5
threshold
0.1
0.2
0.3
0.4
0.6
0.7
0.8
0.9
1.0
PT!EN. As a result, translations in the opposite direction were not sent for
evaluation.</p>
        <p>When compared to the other submissions of the shared task, the rst overview
of the general results con rm that our approach has a good coverage and recall
(LyS-DT achieved the best numbers in this respect), even if the precision is lower
than that of the other participants. In this regard, it must be mentioned that we
decided to submit every output of the systems (even those with 0 con dence), so
that a better analysis of the results should take into account di erent con dence
thresholds. Interestingly, no submission could beat the baselines proposed by the
organisers. The results of our four runs had the same tendencies in each of the
three translations, being LyS-DT the best system followed by LyS-ES, LyS-CA,
and LyS, respectively. Thus, we can infer that using a pivot language worsens
the translation performance. The di erences among each system, however, were
not considerable, and there is no evidence that language family or corpus size
have had a decisive impact, so that only a careful error analysis could shed some
light into these results.</p>
        <p>With regard to each speci c translation pair, LyS-DT was the best submitted
system in the EN!FR direction (0:34 F1), and the second one in FR!PT (0:36
F1) and in PT!EN (0:27 F1), tied and below the Frankfurt team, respectively.
In general, the results of our other systems were slightly below these values.</p>
        <p>We carried out a brief analysis of the LyS-DT submission, aimed at
knowing a little better the output of the system. First, we observed the impact of
the con dence threshold in the nal results. Then, we analyzed the number of
translations of our system, paying special attention to those entries for which
our system were not designed: multiword expressions and proper nouns.</p>
        <p>Figure 1 shows a graph for LyS-DT submission with the evolution of the
four metrics used to measure the performance according to the threshold of the
degree of con dence for translations. As can be seen, the results remain stable
up to a value of 0.5. It is worth noting that the F1 corresponding to a threshold
0 is slightly lower than the F1 between 0.1 and 0.4, so we could have climbed to
the rst position in the nal ranking of participants teams by simply considering
as o cial the results corresponding to a threshold 0.1. From a threshold of 0.5
on there is a rise in accuracy at the cost of a drop in coverage and recall, with
F1 also dropping since the increase in accuracy is not enough to compensate for
the drop in recall. Maximum precision is obtained between thresholds 0.8 and
0.9, but, while for 0.8 the recall has only dropped by half the value for 0.5, from
that point on it descends with a steeper slope.</p>
        <p>25000
20000
s 15000
n
o
tirsan
la 10000
t 5000</p>
        <p>Figure 2 shows the number of translations produced by LyS-DT with each
con dence value. As this gure reveals, most translations have con dence values
between 0.6-0.9. Besides, it is striking that from a total of 191; 373 entries there
were 4; 184 with con dence of 1, and 23; 902 with 0 (which basically means that
no translation was found). Among the rst group, it is worth noting that all
translations with con dence 1 were proper nouns, because they were found in
both pivot and target languages with a high similarity value. However, it is likely
that many of these equivalents are incorrect (see Figure 1), since our
Wikipediabased monolingual models contain proper nouns in di erent languages.</p>
        <p>With regard to the entries with 0 con dence, Table 2 contains the number
of non translated elements discriminated by single or multiword expressions as
well as by PoS tags. For single words, most cases were proper nouns, followed by
common nouns, adverbs, adjectives and verbs. In this respect, it was expected
that our approach could not nd suitable translations for many proper nouns.
Also, several adverbs, adjectives, and nouns could also be wrongly lemmatized
and subsequently not found in our models. However, a brief look at these data
also revealed other issues: on one hand, the input dictionaries contain wrong
lemma PoStag pairs (e.g., actual proper nouns such as BBC, ETA or AT&amp;T
and several adjectives wrongly labeled as nouns). On the other hand, in some
dictionaries there were typos and words in other languages (e.g., the Portuguese
inputs include Spanish words like garant a, asesor, cazador, or ateo, instead of
their Portuguese equivalents garantia, assessor, cacador, and ateu). Therefore,
they were not present in our vocabularies. Apart from these issues, several other
mistranslations of LyS-DT were technical words from speci c domains (e.g.,
Medicine and Biology), which perhaps have low frequency in the Wikipedia
corpus.</p>
        <p>Finally, out of 13; 157 MWEs of our runs, LyS-DT could only translate 344,
proving that our basic MWE approach was not enough to obtain suitable
translations. A simple pre-processing of proper nouns (joint as single tokens) could
improve both their precision and recall. For the other cases, specially for
noncompositional expressions, more complex strategies need to be designed.
4</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Conclusions and future work</title>
      <p>We consider that our participation in the TIAD shared task has been fruitful,
as one of our systems has obtained the rst position among the participant
teams (tied with Frankfurt), while the other three results we sent were placed
from third to fth position. However, our systems failed to beat the \baselines"
established by the organisers. In this regard, we must clarify that these are not
really baselines but quite sophisticated state-of-the-art methods (one based on
multilingual word embeddings and the other one based on the degree of overlap
with the translations obtained by means of a pivot language).</p>
      <p>Besides, it is worth mentioning that our approach only makes use of the
lemma and PoS tag of each entry, so di erent senses of a word are simply inferred
by their distributional contexts. In this respect, it could be interesting to make
use of contextual information for each input word, in order to automatically
select a speci c sense. Also, the use of contextualized distributional models could
be an interesting topic for further research.</p>
      <p>Finally, in future work we plan to improve our method by dealing with
constructions that in the current version are not processed properly. In particular, we
intend to design di erent strategies, both compositional and non-compositional,
for the processing of compound proper nouns as well as for obtaining better
vector representations of multiword expressions.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Artetxe</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Labaka</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Agirre</surname>
          </string-name>
          , E.:
          <article-title>A robust self-learning method for fully unsupervised cross-lingual mappings of word embeddings</article-title>
          .
          <source>In: Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)</source>
          . pp.
          <volume>789</volume>
          {
          <fpage>798</fpage>
          . Association for Computational Linguistics, Melbourne,
          <source>Australia (Jul</source>
          <year>2018</year>
          ), https://www.aclweb.org/anthology/P18-1073
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cardie</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>Unsupervised multilingual word embeddings</article-title>
          .
          <source>In: Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing</source>
          . pp.
          <volume>261</volume>
          {
          <fpage>270</fpage>
          . Association for Computational Linguistics, Brussels, Belgium (
          <year>2018</year>
          ), https://www.aclweb.org/anthology/D18-1024
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Fiser</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sagot</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          :
          <article-title>Constructing a poor man's wordnet in a resource-rich world</article-title>
          .
          <source>Language Resources and Evaluation</source>
          <volume>49</volume>
          (
          <issue>3</issue>
          ),
          <volume>601</volume>
          {635 (Sep
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Fung</surname>
            ,
            <given-names>P.:</given-names>
          </string-name>
          <article-title>A statistical view on bilingual lexicon extraction: From parallel corpora to non-parallel corpora</article-title>
          .
          <source>In: Proceedings of the Third Conference of the Association for Machine Translation in the Americas on Machine Translation and the Information Soup</source>
          . pp.
          <volume>1</volume>
          {
          <fpage>17</fpage>
          . AMTA '
          <volume>98</volume>
          , Springer-Verlag, London, UK (
          <year>1998</year>
          ), http://dl.acm.org/citation.cfm?id=
          <volume>648179</volume>
          .
          <fpage>749226</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Fung</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yee</surname>
            ,
            <given-names>L.Y.</given-names>
          </string-name>
          :
          <article-title>An IR approach for translating new words from nonparallel, comparable texts</article-title>
          .
          <source>In: Proceedings of the 36th Annual Meeting of the Association for Computational Linguistics and 17th International Conference on Computational Linguistics</source>
          , Volume
          <volume>1</volume>
          . pp.
          <volume>414</volume>
          {
          <fpage>420</fpage>
          . Association for Computational Linguistics, Montreal, Quebec,
          <source>Canada (Aug</source>
          <year>1998</year>
          ). https://doi.org/10.3115/980845.980916
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Gamallo</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Garcia</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pineiro</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Martinez-Castan</surname>
            ~o,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pichel</surname>
            ,
            <given-names>J.C.</given-names>
          </string-name>
          :
          <article-title>LinguaKit: a Big Data-based multilingual tool for linguistic analysis and information extraction</article-title>
          .
          <source>In: 2018 Fifth International Conference on Social Networks Analysis, Management and Security (SNAMS)</source>
          . pp.
          <volume>239</volume>
          {
          <fpage>244</fpage>
          .
          <string-name>
            <surname>IEEE</surname>
          </string-name>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Garcia</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gamallo</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Yet Another Suite of Multilingual NLP Tools</article-title>
          . In:
          <article-title>JoseLuis Sierra-Rodr guez and Jose Paulo Leal and Alberto Simo~es (ed.) Languages, Applications and Technologies</article-title>
          .
          <source>Communications in Computer and Information Science</source>
          . pp.
          <volume>65</volume>
          {
          <fpage>75</fpage>
          .
          <source>International Symposium on Languages, Applications and Technologies (SLATE</source>
          <year>2015</year>
          ), Springer (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Habash</surname>
          </string-name>
          , N.:
          <article-title>Introduction to Arabic Natural Language Processing</article-title>
          .
          <source>Synthesis Lectures on Human Language Technologies</source>
          , Morgan &amp; Claypool Publishers (
          <year>2010</year>
          ), http://dx.doi.org/10.2200/S00277ED1V01Y201008HLT010
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Madhyastha</surname>
            ,
            <given-names>P.S.</given-names>
          </string-name>
          ,
          <article-title>Espan~a-</article-title>
          <string-name>
            <surname>Bonet</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>Learning bilingual projections of embeddings for vocabulary expansion in machine translation</article-title>
          .
          <source>In: Proceedings of the 2nd Workshop on Representation Learning for NLP</source>
          . pp.
          <volume>139</volume>
          {
          <fpage>145</fpage>
          . Association for Computational Linguistics, Vancouver, Canada (Aug
          <year>2017</year>
          ). https://doi.org/10.18653/v1/
          <fpage>W17</fpage>
          -2617
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Mikolov</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Corrado</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dean</surname>
            ,
            <given-names>J.:</given-names>
          </string-name>
          <article-title>E cient estimation of word representations in vector space</article-title>
          .
          <source>In: Workshop Proceedings of the International Conference on Learning Representations (ICLR)</source>
          <year>2013</year>
          (
          <year>2013</year>
          ), arXiv preprint arXiv:
          <volume>1301</volume>
          .
          <fpage>3781</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Padro</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Stanilovsky</surname>
          </string-name>
          , E.:
          <article-title>Freeling 3.0: Towards wider multilinguality</article-title>
          .
          <source>In: Proceedings of the Eighth International Conference on Language Resources</source>
          and
          <article-title>Evaluation (LREC-</article-title>
          <year>2012</year>
          ). pp.
          <volume>2473</volume>
          {
          <fpage>2479</fpage>
          .
          <string-name>
            <surname>European Language Resources Association</surname>
          </string-name>
          (ELRA), Istanbul, Turkey (May
          <year>2012</year>
          ), http://www.lrecconf.org/proceedings/lrec2012/pdf/430 Paper.pdf
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Poibeau</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Saggion</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Piskorski</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yangarber</surname>
          </string-name>
          , R.: Multi-source,
          <source>Multilingual Information Extraction and Summarization</source>
          . Springer Publishing Company, Incorporated (
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Prochasson</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fung</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Rare word translation extraction from aligned comparable documents</article-title>
          .
          <source>In: Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies</source>
          . pp.
          <volume>1327</volume>
          {
          <fpage>1335</fpage>
          . Association for Computational Linguistics, Portland, Oregon, USA (Jun
          <year>2011</year>
          ), https://www.aclweb.org/anthology/P11-1133
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Rapp</surname>
            ,
            <given-names>R.:</given-names>
          </string-name>
          <article-title>A methodology for bilingual lexicon extraction from comparable corpora</article-title>
          .
          <source>In: Proceedings of the Fourth Workshop on Hybrid Approaches to Translation (HyTra)</source>
          . pp.
          <volume>46</volume>
          {
          <fpage>55</fpage>
          . Association for Computational Linguistics, Beijing (Jul
          <year>2015</year>
          ). https://doi.org/10.18653/v1/
          <fpage>W15</fpage>
          -4108
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Rapp</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Xu</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zock</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sharo</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Forsyth</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Babych</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chu</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nakazawa</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kurohashi</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>New areas of application of comparable corpora</article-title>
          . In: Skadina,
          <string-name>
            <given-names>I.</given-names>
            ,
            <surname>Gaizauskas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            ,
            <surname>Babych</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            ,
            <surname>Ljubesic</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            ,
            <surname>Tu s</surname>
          </string-name>
          , D.,
          <string-name>
            <surname>Vasiljevs</surname>
            ,
            <given-names>A</given-names>
          </string-name>
          . (eds.)
          <article-title>Using Comparable Corpora for Under-Resourced Areas of Machine Translation</article-title>
          , pp.
          <volume>255</volume>
          {
          <fpage>290</fpage>
          . Springer, Cham (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <surname>Rehurek</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sojka</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Software Framework for Topic Modelling with Large Corpora</article-title>
          .
          <source>In: Proceedings of the LREC 2010 Workshop on New Challenges for NLP Frameworks</source>
          . pp.
          <volume>45</volume>
          {
          <fpage>50</fpage>
          .
          <string-name>
            <surname>European Language Resources Association</surname>
          </string-name>
          (ELRA), Valletta, Malta (May
          <year>2010</year>
          ), http://is.muni.cz/publication/884893/en
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          17.
          <string-name>
            <surname>Steinberger</surname>
            ,
            <given-names>R.:</given-names>
          </string-name>
          <article-title>A survey of methods to easy the development of highly multilingual text mining applications</article-title>
          .
          <source>Language Resources and Evaluation</source>
          <volume>46</volume>
          ,
          <issue>155</issue>
          {
          <fpage>176</fpage>
          (
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          18.
          <string-name>
            <surname>Vilares</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gomez-Rodr guez</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Alonso</surname>
            ,
            <given-names>M.A.</given-names>
          </string-name>
          :
          <article-title>Universal, unsupervised (rulebased), uncovered sentiment analysis</article-title>
          .
          <source>Knowledge-Based Systems 118</source>
          ,
          <fpage>45</fpage>
          {
          <fpage>55</fpage>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          19.
          <string-name>
            <surname>Wong</surname>
            ,
            <given-names>K.F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Xu</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          , sheng Zhang,
          <string-name>
            <surname>Z.</surname>
          </string-name>
          : Introduction to Chinese
          <source>Natural Language Processing, Synthesis Lectures on Human Language Technologies</source>
          , vol.
          <volume>4</volume>
          . Morgan &amp; Claypool Publishers, San Rafael, CA (
          <year>2010</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>