<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Are Crescia and Piadina the Same? Towards Identifying Synonymy or Non-Synonymy between Italian Words to Enable Crowdsourcing from Language Learners</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Lavinia Aparaschivei</string-name>
          <email>lavinianicoleta.aparaschivei@eurac.edu</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Lionel Nicolas</string-name>
          <email>lionel.nicolas@eurac.edu</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Alberto Barr o´n-Ceden˜ o</string-name>
          <email>a.barron@unibo.it</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>DIT-Universita` di Bologna</institution>
          ,
          <addr-line>Forl`ı</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Institute for Applied Linguistics, Eurac Research</institution>
          ,
          <addr-line>Bolzano</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>We introduce a method to generate candidate pairs of related Italian words sharing (or not) synonymous relations from the ConceptNet knowledgebase. The pairs are intended to generate questions for a vocabulary trainer which combines exercises to enhance vocabulary skills with the implicit crowdsourcing of linguistic knowledge about the semantic relations between words. Our method relies on the idea that pairs of synonyms in a language tend to translate to pairs of synonyms in other languages. We generated 85k candidate pairs of Italian synonyms that can be used to produce questions for both teaching (3.8k pairs) and crowdsourcing purposes (80k pairs). Follow-up efforts are however needed in order to generate a complementary set of questions.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1 Introduction</title>
      <p>Our efforts target the automatic generation of
semantically-related candidate pairs of Italian
words with a focus on synonymy. We address a
cold start issue for a vocabulary trainer
combining exercises to enhance vocabulary skills with
the implicit crowdsourcing of linguistic
knowledge about the semantic relations between words.</p>
      <p>While targeting a specific use case, our method
contributes to a larger effort aimed at narrowing
gaps on two fronts. On the NLP front, over the
past few decades varied efforts have targeted the
efficient creation, extension, and maintenance of</p>
      <p>
        Copyright © 2021 for this paper by its authors. Use
permitted under Creative Commons License Attribution 4.0
International (CC BY 4.0).
resources, including crowdsourcing through
platforms such as Amazon Mechanical Turk
        <xref ref-type="bibr" rid="ref15 ref4 ref6 ref6">(Ganbold et al., 2018; Potthast et al., 2018)</xref>
        . Still, the
subject remains an open issue. On the
computerassisted language learning (CALL) front, the
automatic generation of exercise content from NLP
resources is almost non-existent, despite the fact that
some of these datasets encode the knowledge that
learners are often tested on (e.g., lexical
knowledge). This absence is probably due to
differences in expectations with respect to linguistic
accuracy: learning materials are usually close to
perfect, whereas NLP resources rarely are.
Generating content from imperfect datasets poses a
challenge in terms of its suitability for learning.
      </p>
      <p>
        We contribute to narrowing these gaps by
producing data to tackle a cold start issue for a
vocabulary trainer designed to both teach language
and crowdsource linguistic knowledge from
learners. We generate a collection of candidate pairs of
Italian words tied to confidence scores, allowing to
decide which pairs should be used for learning or
for crowdsourcing purposes. Our method projects
synonymy information in ConceptNet
        <xref ref-type="bibr" rid="ref20">(Speer et
al., 2017)</xref>
        from non-Italian onto Italian words. The
obtained results show that we adequately tackle
part of the cold start issue, while follow-up efforts
are needed to address the remaining part.
      </p>
      <p>The rest of the paper is organised as follows:
Section 2 discusses the specific purpose of our
method. Section 3 summarises related work.
Section 4 and Section 5 describe how the candidate
pairs are generated and scored. Finally, Section 6
discusses how suitable the pairs are for our specific
use case and Section 7 provides closing remarks.
2</p>
    </sec>
    <sec id="sec-2">
      <title>Background</title>
      <p>
        Ours and previous related work
        <xref ref-type="bibr" rid="ref12 ref16 ref16 ref17 ref6 ref6 ref8 ref8">(Lyding et al.,
2019; Rodosthenous et al., 2019; Rodosthenous
et al., 2020; Nicolas et al., 2021)</xref>
        all contribute to
a wider effort to research an implicit
crowdsourcing paradigm built upon the idea that the curation
of NLP resources and language learning are
sibling endeavors
        <xref ref-type="bibr" rid="ref11">(Nicolas et al., 2020)</xref>
        . On the one
hand, (NLP) researchers try to create models to
“teach” a computer to process and/or produce
language utterances. On the other hand, learners
create a model, in the form of personal knowledge,
to process and/or produce language utterances too.
Allowing learners to express their knowledge can
contribute to enhance an NLP resource, under
specific conditions. This paradigm substitutes the
expert manpower typically required to curate NLP
resources with a non-expert crowd of learners.
      </p>
      <p>Expert manpower can be substituted by
nonexpert crowds, as exemplified by the numerous
efforts to use Amazon Mechanical Turk (AMT) to
build NLP datasets; e.g., Ganbold et al. (2018)
or Potthast et al. (2018). The synergy exploited
by this paradigm between NLP and CALL is
particularly interesting. Indeed, NLP can be used
to enhance CALL methods and, as such, it can
grant an intrinsic added value for the target crowd
that is not limited by any type of resource,
unlike other crowdsourcing approaches relying on
extrinsic added values (e.g. monetary incentives
in AMT). In addition, from the NLP
perspective, the crowd of learners that could potentially
be reached is immense. Accordingly, an
unprecedented amount of data could theoretically be
crowdsourced by exploiting such a synergy.</p>
      <p>Nicolas et al. (2021) showed that linguistic
knowledge about an entry of an NLP dataset can
be obtained at expert quality level, provided that
the same judgement is asked to a sufficient
number of learners. This is mostly true when simple
Boolean questions are used. Even linguistic
judgements of inferior reliability (e.g. 70%) contribute
to approaching statistical certainty about the right
answer to Boolean questions.1 This approach
favors quantity over quality to meet its goals and can
be used to produce new entries or to validate the
existing ones in an NLP dataset.</p>
      <p>
        V-trel is a vocabulary trainer that implements
this paradigm to teach and crowdsource
knowledge on semantic relations between words
        <xref ref-type="bibr" rid="ref12 ref16 ref16 ref6 ref6 ref8 ref8">(Lyding
et al., 2019; Rodosthenous et al., 2019; Nicolas
et al., 2021)</xref>
        . V-trel includes two types of
questions. Open questions ask for words sharing a
spe1While the results obtained tend to confirm the viability
of the approach, many aspects remain unexplored; e.g., the
difficulty of a question in the aggregation process.
cific relation with a given one (e.g., “give me a
synonym of x”). Closed questions show a pair of
words and ask the Boolean question of whether
they share a specific relation (e.g., “are x and y
synonyms?”). From the crowdsourcing
perspective, open questions are mostly intended to
crowdsource additional knowledge (i.e. to crowdsource
new candidate entries), whereas closed questions
are designed to crowdsource judgements on the
knowledge suggested in the open questions or
already encoded in ConceptNet (i.e. to validate
existing entries or new candidate entries).
      </p>
      <p>
        As empirically observed, closed questions
should elicit positive and negative answers from
learners
        <xref ref-type="bibr" rid="ref17">(Rodosthenous et al., 2020)</xref>
        . Otherwise,
when learners understand that the trainer tends to
continuously expect the same answer (e.g., “yes”),
they tend to give the same default answer
mechanically, without producing meaningful judgments.
      </p>
      <p>We aim at generating closed questions
expecting both types of answers. We need to identify
pairs of synonyms to produce questions eliciting a
positive answer and word pairs sharing a semantic
relation other than synonymy (e.g. antonymy) to
elicit negative answers. We refer to them as
nonsynonyms in the rest of the paper. To elicit positive
answers, we use synonyms, such as “house” and
“home”. To elicit negative answers, we need
nonsynonyms such as “good” and “bad”. It is worth
noting that we do not consider as non-synonyms
pairs of unrelated words such as ”house” and
”dog”’ because they do not share any kind of
semantic relation. Questions eliciting negative
answers generated from them would be of poor
quality and would not pose any challenge to learners.</p>
      <p>Since v-trel favours teaching over
crowdsourcing to maintain its pool of users, closed questions
designed to crowdsource knowledge from learners
should be served on a low frequency. This implies
the need to decide which questions can be used for
teaching and which for crowdsourcing purposes.
Hence, the need of a confidence score to divide
the questions into the two sets.</p>
      <p>Our method aims at replacing the closed
questions generated from ConceptNet, whose expected
answers have quality issues since ConceptNet is,
as most NLP resources, an imperfect dataset and
for which we cannot tell apart the questions that
can be used for teaching and for crowdsourcing
purposes. Even though the aggregation of
answers crowdsourced from learners would solve
these problems, we have a cold start issue similar
to the chicken and egg paradox: the issues
cannot be solved without offering the tool but the tool
cannot be offered without solving the issues first.
3</p>
    </sec>
    <sec id="sec-3">
      <title>Related Work</title>
      <p>
        With respect to the automatic generation of
language learning exercises, only little automatic
generation is performed directly from NLP
resources so far. Most efforts focus on exercises
known as a “cloze” (deletion) test, where learners
have to fill word gaps in a text
        <xref ref-type="bibr" rid="ref5 ref6 ref6">(Hill and Simha,
2016; Katinskaia et al., 2018; Lee et al., 2019)</xref>
        .
The literature from the last four editions of the
toptwo venues concerned with using NLP for CALL2
confirms that current efforts, aside from ours, are
mostly dedicated to the generation of cloze
exercises
        <xref ref-type="bibr" rid="ref10 ref19">(Santhi Ponnusamy and Meurers, 2021)</xref>
        ,
the modelling of the learner knowledge
        <xref ref-type="bibr" rid="ref1">(Araneta
et al., 2020)</xref>
        , or the detection and/or correction of
mistakes in written text
        <xref ref-type="bibr" rid="ref21">(U¨ ksik et al., 2021)</xref>
        . Some
preliminary efforts exist on the automatic
generation of exercises from Finnish and Hungarian NLP
resources.3 Despite the relatively narrow nature
of the exercises we aim at generating4, our work
represents one of the few efforts targeting the
automatic generation of language learning exercises
from NLP resources.
      </p>
      <p>
        Since our method generates pairs of synonyms
from an existing knowledge base, it shares
common ground with approaches to build or extend
similar datasets. In that respect, the state of the
art is mostly concerned with the creation and
curation of WordNets for which various semi- and
fully-automatic techniques have been developed,
especially for languages other than English.
Following Vossen (1996), these methods can be
categorised as using either a merge or an
expansion approach or both. The merge approach
employs monolingual resources to create a
standalone WordNet and was adopted for
EuroWordNet
        <xref ref-type="bibr" rid="ref23">(Vossen, 1998)</xref>
        , the Polish WordNet
        <xref ref-type="bibr" rid="ref2">(Derwojedowa et al., 2008)</xref>
        , the Norwegian
WordNet
        <xref ref-type="bibr" rid="ref3">(Fjeld and Nygaard, 2009)</xref>
        and the Danish
WordNet
        <xref ref-type="bibr" rid="ref13">(Pedersen et al., 2009)</xref>
        . The expansion
2The BEA Workshop https://aclanthology.o
rg/venues/bea/, and the NLP4CALL Workshop http
s://aclanthology.org/venues/nlp4call/
3See the following PhD project: https://spraakba
nken.gu.se/cms/sites/default/files/2021/
nlp4call2021 researchnotes1 talk1.pdf
4It would certainly be interesting to extend such exercises
with a sentence context.
approach uses a source WordNet and translates
its synsets into the target language. It was used
to build MultiWordNet
        <xref ref-type="bibr" rid="ref14">(Pianta et al., 2002)</xref>
        , the
Finnish WordNet
        <xref ref-type="bibr" rid="ref7">(Linden and Carlson, 2010)</xref>
        , the
French WordNet WOLF
        <xref ref-type="bibr" rid="ref18">(Sagot and Fisˇer, 2008)</xref>
        ,
and to enhance a Persian WordNet
        <xref ref-type="bibr" rid="ref10 ref19 ref20 ref9">(Mousavi and
Faili, 2017; Mousavi and Faili, 2021)</xref>
        .
      </p>
      <p>Our method employs an expansion approach:
it projects knowledge from other languages onto
Italian, but it differs in three aspects. First, it
relies on a different type of dataset: ConceptNet.5
Second, the output is not a final product, but a
“raw” dataset to be polished by crowdsourcing.
Third, it aims at identifying both synonyms and
non-synonyms, whereas the aforementioned
methods are mostly concerned with synonyms only.
4</p>
    </sec>
    <sec id="sec-4">
      <title>Generating Candidate Pairs</title>
      <p>Our hypothesis is that if two non-Italian words
are marked as synonyms in ConceptNet and such
words are translations of a pair of Italian words,
then the Italian words are synonyms with a high
likelihood. For instance, the pair {house, home}
in English with respect to {casa, abitazione} in
Italian. The greater the number of such pairs of
non-Italian words are identified (e.g., {maison,
logement} in French, {casa, vivienda} in Spanish),
the more likely the Italian words are to be
synonymous. Hereafter, we refer to the number of pairs
of non-Italian words projected onto an Italian pair
as Nb-projected-syn-pairs.</p>
      <p>At the same time, we assumed that the
incorrect candidate pairs of synonyms generated would
mostly constitute a valid set of candidate pairs
of non-synonyms. As such, most candidate pairs
would be used to tackle our specific use case.</p>
      <p>We used this logic for all languages available in
ConceptNet. In order to seamlessly add the data
already available on Italian synonyms, we
considered the Italian part of ConceptNet as describing
just another non-Italian language. Hence, we
considered all Italian words as translations of
themselves in this “extra” language.</p>
      <p>We extracted 84, 602 candidate pairs of Italian
synonyms and randomly sampled and evaluated
a subset of 1, 120 pairs to build a gold standard.</p>
      <p>5ConceptNet is a multilingual knowledge base that
represents commonly-used words and phrases as well as the
relationships between them. It currently holds more than
34 million assertions about words: terma &lt;relation&gt;
termb. ConceptNet can be accessed via an API, making it
easy to integrate into applications.</p>
      <p>The annotation procedure started by reflecting the
information in well-known online Italian
dictionaries: Treccani, De Mauro, Gabrielli,
SabatiniColetti, Rizzoli, and Virgilio.6 When a candidate
pair was not found in these dictionaries, an
annotator studied the definitions of the two words and
searched for a third word referenced as a synonym
of both words in the pair. We only kept instances
where the annotator showed a high confidence. In
total, 515 were labeled as correct pairs and 485
as incorrect. We discarded 120 pairs. From the
1, 000 annotated instances, 403 directly reflect the
information of reference dictionaries, whereas 597
reflect the stand of the annotator.</p>
      <p>By extrapolating the ratio observed in the gold
standard, we estimate that 51.2% of the candidate
pairs (∼ 43.4k pairs) are indeed synonyms. In
comparison, 19,906 Italian word pairs are marked
as synonyms in ConceptNet. We randomly
sampled and annotated 200 of them with the procedure
used to build the gold standard. Our estimation
that 84% of them (∼ 16.7k pairs) are valid. Our
set of candidate pairs of Italian synonyms is thus
larger, but has lower quality. Using these pairs
directly to generate closed questions eliciting a
positive answer would thus defeat our goal of
improving the quality of the closed questions.
5</p>
    </sec>
    <sec id="sec-5">
      <title>Computing Confidence Scores</title>
      <p>We aim at discriminating between instances
intended to generate questions eliciting positive and
negative answers, while discriminating questions
used for teaching or crowdsourcing purposes. We
relied on a binary classifier to flag candidate pairs
as correct and incorrect instances of synonyms.
The predictions are used to decide on the kind
of answers to elicit —candidate pairs predicted
as correct are used to generate questions
eliciting a positive answer and vice-versa. The
associated confidence scores are used to discriminate
between questions used for teaching and for
crowdsourcing. We used the aforementioned gold
standard to train the classifier.</p>
      <p>The features are the following. (1) The
aforementioned Nb-projected-syn-pairs for each pair.</p>
      <p>6https://www.treccani.it; https://dizi
onario.internazionale.it; https://www.gr
andidizionari.it/Dizionario Italiano/; ht
tps://dizionari.corriere.it/dizionario
italiano/; https://dizionari.corriere.i
t/dizionario sinonimi contrari/; https:
//sapere.virgilio.it.
model
Random forest
Logistic regression
Random Tree
Baseline</p>
      <p>F1
67.0
69.2
65.9
67.7
(2) To distinguish the languages from which the
projection of knowledge happened, we computed
per language the size of each subset of pairs of
non-Italian words projected onto the candidate
pair (which, together, sum up to
Nb-projectedsyn-pairs). (3) To express “relatedness”, we
computed the size of the set of non-Italian pairs of
words both marked as sharing a semantic relation
(i.e. not only synonymy) and as translations of
the candidate pair. We refer hereafter to this
number as Nb-projected-all-pairs. We also computed a
ratio obtained by dividing Nb-projected-syn-pairs
by Nb-projected-all-pairs. (4) To indicate if a
candidate pair might be better suited to another
semantic relation, we computed per semantic
relation the size of the subsets of non-Italian pairs
of words both marked as sharing a semantic
relation other than synonymy and as translations of
the candidate pair, as well as a ratio value by
dividing these sizes by Nb-projected-syn-pairs and
a difference by subtracting Nb-projected-syn-pairs
to them. (5) A last set of features represents the
most found relation, besides synonymy, in these
non-Italian pairs of words (i.e. the top
“competitor”) by providing its type and duplicating the
corresponding size of the subset of non-Italian pairs
of words, ratio, and difference.</p>
      <p>Since our gold standard is small, we ran a
leaveone-out cross validation process to assess the
quality of the predictions for a number of classifiers
with default settings.7 Table 1 shows the
performance obtained by three of them plus a baseline
that labels all pairs as correct. Even if the
logistic regressor obtains the highest F1, we adopt the
model with the highest precision: the random
forest. The reason is that we have observed
empirically that precision is the most adequate indicator
of how much the confidence scores would
corre7We used Weka 3.8.8; https://www.cs.waikato
.ac.nz/ml/weka/.
100
80
n
o
ii
s
c
reP 60
40
ito 0.50
a
R
1.00
late with the quality of the predicted labels.8
6</p>
    </sec>
    <sec id="sec-6">
      <title>Categorising Candidate Pairs</title>
      <p>Once the binary classification was completed, we
had to distinguish which pairs could be used to
generate teaching and which to generate
crowdsourcing questions. For that, we studied the
correlation between confidence scores and quality of
prediction.</p>
      <p>Figure 1 shows the precision obtained when
thresholding at different confidence score values.
The “all” curve shows a clear correlation between
the quality of the label predicted and the
confidence scores, which was the main result we were
aiming for. However, the performance differs
noticeably with respect to the label predicted: the
curve associated with pairs predicted as “correct”
grows as expected, whereas the one for pairs
predicted as “incorrect” does not. The reason can be
observed through the ratio of labels predicted
according to confidence scores.</p>
      <p>As Figure 2 shows, label “incorrect” was rarely
predicted with high confidence scores. This is
because our method is inherently oriented towards
identifying pairs of synonyms. Accordingly, the
pairs outputted that are not synonyms are also not,
as we hoped for, pairs of non-synonyms. They
are mostly random noise induced by homonyms
in other languages. For example, the
candidate {fuoco, licenziare} was generated
because the English words {fire, dismiss}
are synonyms. Fire has several homonyms with
different senses, one of which translates to fuoco
in Italian. Our set of candidate pairs thus contains
8Future efforts will explore more direct and quantifiable
means of formally informing this selection; cf. Section 7.
only a few non-synonym pairs that the binary
classifier struggles to spot. Therefore, our method
cannot be used at present to generate closed questions
eliciting negative answers.</p>
      <p>This is not the case for pairs predicted as
correct. For example, by using a minimum threshold
of 0.996 on the confidence scores, we can select
3,829 pairs for which the predicted “correct” label
is 94.44% reliable. This represents a set of pairs
of reasonable size and better quality than the ones
encoded in ConceptNet, which allow us to address
part of the cold-start issue.
7</p>
    </sec>
    <sec id="sec-7">
      <title>Conclusions and Ongoing Work</title>
      <p>We presented a method to generate candidate
pairs of Italian words that are synonyms or
nonsynonyms of one another from ConceptNet. These
pairs will be used to generate questions used by a
vocabulary trainer designed to combine the
crowdsourcing of NLP datasets with language learning.
While overtime all questions will be used for both
teaching and crowdsourcing purposes, part of the
pairs generated will at first be used to teach
learners while the other part will at first be used to
crowdsource knowledge in order to enhance
ConceptNet. The obtained pairs, known to be correct
synonyms in advance, can be served to the
learners to improve their vocabulary skills. Another
subset, whose correctness is still to be confirmed,
can be served to the learners for validation and to
decide whether the synonym connection between
them should be added to ConceptNet or not.</p>
      <p>Our results show that we can produce adequate
data to generate part of the questions, while we
are still unable to produce the data required to
generate the complementary set of questions. In
order to tackle the latter, we are devising a
similar approach to identify candidate pairs of
nonsynonyms. We are adapting our overall procedure
for the pairs of Italian words marked as
translations of non-Italian words sharing any semantic
relations (e.g. antonyms or hyponyms) instead of
only considering the ones marked as translations
of non-Italian words sharing a synonymy relation.</p>
      <p>We are also interested in exploring possibilities
to perform a more informed selection of the binary
classification algorithm and will explore metrics to
quantify the correlation between confidence scores
and the quality of the predicted labels (e.g.
Pearson, Kendall). In the future, we aim at running a
crowdsourcing experiment with students of Italian
as a second language with the produced data.
the Eleventh International Conference on Language
Resources and Evaluation (LREC 2018).</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <string-name>
            <given-names>Marianne</given-names>
            <surname>Grace</surname>
          </string-name>
          <string-name>
            <given-names>Araneta</given-names>
            , Gu¨ls¸en Eryig˘it, Alexander Ko¨nig,
            <surname>Ji-Ung</surname>
          </string-name>
          <string-name>
            <surname>Lee</surname>
          </string-name>
          , Ana Lu´ıs, Verena Lyding, Lionel Nicolas, Christos Rodosthenous, and
          <string-name>
            <given-names>Federico</given-names>
            <surname>Sangati</surname>
          </string-name>
          .
          <year>2020</year>
          .
          <article-title>Substituto - a synchronous educational language game for simultaneous teaching and crowdsourcing</article-title>
          .
          <source>In Proceedings of the 9th Workshop on NLP for Computer Assisted Language Learning</source>
          , pages
          <fpage>1</fpage>
          -
          <lpage>9</lpage>
          , Gothenburg, Sweden, November. LiU Electronic Press.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          <string-name>
            <given-names>Magdalena</given-names>
            <surname>Derwojedowa</surname>
          </string-name>
          , Maciej Piasecki, Stanisªaw Szpakowicz, Magdalena Zawisławska, and
          <string-name>
            <given-names>Bartosz</given-names>
            <surname>Broda</surname>
          </string-name>
          .
          <year>2008</year>
          .
          <article-title>Words, concepts and relations in the construction of polish wordnet</article-title>
          .
          <source>Proceedings of GWC 2008</source>
          , pages
          <fpage>162</fpage>
          -
          <lpage>177</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          <string-name>
            <given-names>Ruth</given-names>
            <surname>Vatvedt</surname>
          </string-name>
          Fjeld and
          <string-name>
            <given-names>Lars</given-names>
            <surname>Nygaard</surname>
          </string-name>
          .
          <year>2009</year>
          .
          <article-title>Norneta monolingual wordnet of modern norwegian</article-title>
          .
          <source>In NODALIDA 2009 workshop: WordNets and other Lexical Semantic Resources-between Lexical Semantics, Lexicography, Terminology and Formal Ontologies</source>
          , volume
          <volume>7</volume>
          , pages
          <fpage>13</fpage>
          -
          <lpage>16</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          <string-name>
            <given-names>Amarsanaa</given-names>
            <surname>Ganbold</surname>
          </string-name>
          , Altangerel Chagnaa, and Ga´bor Bella.
          <year>2018</year>
          .
          <article-title>Using crowd agreement for wordnet localization</article-title>
          .
          <source>In Proceedings of the Eleventh International Conference on Language Resources and Evaluation (LREC</source>
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          <string-name>
            <given-names>Jennifer</given-names>
            <surname>Hill</surname>
          </string-name>
          and
          <string-name>
            <given-names>Rahul</given-names>
            <surname>Simha</surname>
          </string-name>
          .
          <year>2016</year>
          .
          <article-title>Automatic generation of context-based fill-in-the-blank exercises using co-occurrence likelihoods and google ngrams</article-title>
          .
          <source>In Proceedings of the 11th Workshop on Innovative Use of NLP for Building Educational Applications</source>
          , pages
          <fpage>23</fpage>
          -
          <lpage>30</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          <string-name>
            <given-names>Anisia</given-names>
            <surname>Katinskaia</surname>
          </string-name>
          , Javad Nouri,
          <string-name>
            <given-names>Roman</given-names>
            <surname>Yangarber</surname>
          </string-name>
          , et al.
          <year>2018</year>
          .
          <article-title>Revita: a language-learning platform at the intersection of its and call</article-title>
          .
          <source>In Proceedings of Ji-Ung Lee</source>
          ,
          <string-name>
            <given-names>Erik</given-names>
            <surname>Schwan</surname>
          </string-name>
          , and Christian M Meyer.
          <year>2019</year>
          .
          <article-title>Manipulating the difficulty of c-tests</article-title>
          . arXiv preprint arXiv:
          <year>1906</year>
          .06905.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          <string-name>
            <given-names>K</given-names>
            <surname>Linden and L Carlson</surname>
          </string-name>
          .
          <year>2010</year>
          .
          <article-title>Construction of a finnwordnet</article-title>
          .
          <source>Nordic Journal of Lexicography</source>
          ,
          <volume>17</volume>
          :
          <fpage>119</fpage>
          -
          <lpage>140</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          <string-name>
            <given-names>Verena</given-names>
            <surname>Lyding</surname>
          </string-name>
          , Christos Rodosthenous, Federico Sangati, Umair ul Hassan, Lionel Nicolas, Alexander Ko¨nig, Jolita Horbacauskiene, and
          <string-name>
            <given-names>Anisia</given-names>
            <surname>Katinskaia</surname>
          </string-name>
          .
          <year>2019</year>
          .
          <article-title>v-trel: Vocabulary trainer for tracing word relations - an implicit crowdsourcing approach</article-title>
          .
          <source>In Proceedings of the International Conference on Recent Advances in Natural Language Processing (RANLP</source>
          <year>2019</year>
          ), pages
          <fpage>674</fpage>
          -
          <lpage>683</lpage>
          , Varna, Bulgaria, September. INCOMA Ltd.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          <string-name>
            <given-names>Zahra</given-names>
            <surname>Mousavi</surname>
          </string-name>
          and
          <string-name>
            <given-names>Heshaam</given-names>
            <surname>Faili</surname>
          </string-name>
          .
          <year>2017</year>
          .
          <article-title>Persian wordnet construction using supervised learning</article-title>
          .
          <source>arXiv preprint arXiv:1704</source>
          .
          <fpage>03223</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          <string-name>
            <given-names>Zahra</given-names>
            <surname>Mousavi</surname>
          </string-name>
          and
          <string-name>
            <given-names>Heshaam</given-names>
            <surname>Faili</surname>
          </string-name>
          .
          <year>2021</year>
          .
          <article-title>Developing the persian wordnet of verbs using supervised learning</article-title>
          .
          <source>Transactions on Asian and Low-Resource Language Information Processing</source>
          ,
          <volume>20</volume>
          (
          <issue>4</issue>
          ):
          <fpage>1</fpage>
          -
          <lpage>18</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          <string-name>
            <given-names>Lionel</given-names>
            <surname>Nicolas</surname>
          </string-name>
          , Verena Lyding, Claudia Borg, Corina Fora˘scu, Kare¨n Fort, Katerina Zdravkova, Iztok Kosem, Jaka Cˇ ibej,
          <source>Sˇ pela Arhar Holdt</source>
          ,
          <string-name>
            <given-names>Alice</given-names>
            <surname>Millour</surname>
          </string-name>
          , et al.
          <year>2020</year>
          .
          <article-title>Creating expert knowledge by relying on language learners: a generic approach for mass-producing language resources by combining implicit crowdsourcing and language learning</article-title>
          .
          <source>In Proceedings of The 12th Language Resources and Evaluation Conference</source>
          , pages
          <fpage>268</fpage>
          -
          <lpage>278</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          <string-name>
            <given-names>Lionel</given-names>
            <surname>Nicolas</surname>
          </string-name>
          , Lavinia Nicoleta Aparaschivei, Verena Lyding, Christos Rodosthenous, Federico Sangati, Alexander Ko¨nig, and
          <string-name>
            <given-names>Corina</given-names>
            <surname>Forascu</surname>
          </string-name>
          .
          <year>2021</year>
          .
          <article-title>An experiment on implicitly crowdsourcing expert knowledge about Romanian synonyms from language learners</article-title>
          .
          <source>In Proceedings of the 10th Workshop on NLP for Computer Assisted Language Learning</source>
          , pages
          <fpage>1</fpage>
          -
          <lpage>14</lpage>
          , Online, May. LiU Electronic Press.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          <string-name>
            <given-names>Bolette</given-names>
            <surname>Sandford</surname>
          </string-name>
          <string-name>
            <surname>Pedersen</surname>
          </string-name>
          , Sanni Nimb, Jørg Asmussen, Nicolai Hartvig Sørensen, Lars TrapJensen, and
          <string-name>
            <given-names>Henrik</given-names>
            <surname>Lorentzen</surname>
          </string-name>
          .
          <year>2009</year>
          .
          <article-title>Dannet: the challenge of compiling a wordnet for danish by reusing a monolingual dictionary</article-title>
          .
          <source>Language resources and evaluation</source>
          ,
          <volume>43</volume>
          (
          <issue>3</issue>
          ):
          <fpage>269</fpage>
          -
          <lpage>299</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          <string-name>
            <given-names>Emanuele</given-names>
            <surname>Pianta</surname>
          </string-name>
          , Luisa Bentivogli, and
          <string-name>
            <given-names>Christian</given-names>
            <surname>Girardi</surname>
          </string-name>
          .
          <year>2002</year>
          .
          <article-title>Multiwordnet: developing an aligned multilingual database</article-title>
          .
          <source>In First international conference on global WordNet</source>
          , pages
          <fpage>293</fpage>
          -
          <lpage>302</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          <string-name>
            <given-names>Martin</given-names>
            <surname>Potthast</surname>
          </string-name>
          , Tim Gollub, Kristof Komlossy,
          <string-name>
            <given-names>Sebastian</given-names>
            <surname>Schuster</surname>
          </string-name>
          , Matti Wiegmann, Erika Patricia Garces Fernandez, Matthias Hagen, and
          <string-name>
            <given-names>Benno</given-names>
            <surname>Stein</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>Crowdsourcing a large corpus of clickbait on twitter</article-title>
          .
          <source>In Proceedings of the 27th international conference on computational linguistics</source>
          , pages
          <fpage>1498</fpage>
          -
          <lpage>1507</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          <string-name>
            <surname>Christos</surname>
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Rodosthenous</surname>
            , Verena Lyding, Alexander Koenig, Jolita Horbacauskiene, Anisia Katinskaia, Umair ul Hassan, Nicos Isaak, Federico Sangati, and
            <given-names>Lionel</given-names>
          </string-name>
          <string-name>
            <surname>Nicolas</surname>
          </string-name>
          .
          <year>2019</year>
          .
          <article-title>Designing a prototype architecture for crowdsourcing language resources</article-title>
          .
          <source>In LDK.</source>
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          <string-name>
            <given-names>Christos</given-names>
            <surname>Rodosthenous</surname>
          </string-name>
          , Verena Lyding, Federico Sangati, Alexander Ko¨nig, Umair ul Hassan, Lionel Nicolas, Jolita Horbacauskiene, Anisia Katinskaia, and
          <string-name>
            <given-names>Lavinia</given-names>
            <surname>Aparaschivei</surname>
          </string-name>
          .
          <year>2020</year>
          .
          <article-title>Using crowdsourced exercises for vocabulary training to expand conceptnet</article-title>
          .
          <source>In Proceedings of The 12th Language Resources and Evaluation Conference</source>
          , pages
          <fpage>307</fpage>
          -
          <lpage>316</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          <string-name>
            <given-names>Benoˆıt</given-names>
            <surname>Sagot and Darja Fisˇer</surname>
          </string-name>
          .
          <year>2008</year>
          .
          <article-title>Building a free french wordnet from multilingual resources</article-title>
          .
          <source>In OntoLex.</source>
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          <string-name>
            <given-names>Haemanth</given-names>
            <surname>Santhi</surname>
          </string-name>
          Ponnusamy and
          <string-name>
            <given-names>Detmar</given-names>
            <surname>Meurers</surname>
          </string-name>
          .
          <year>2021</year>
          .
          <article-title>Employing distributional semantics to organize task-focused vocabulary learning</article-title>
          .
          <source>In Proceedings of the 16th Workshop on Innovative Use of NLP for Building Educational Applications</source>
          , pages
          <fpage>26</fpage>
          -
          <lpage>36</lpage>
          , Online, April. Association for Computational Linguistics.
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          <string-name>
            <given-names>Robyn</given-names>
            <surname>Speer</surname>
          </string-name>
          , Joshua Chin, and
          <string-name>
            <given-names>Catherine</given-names>
            <surname>Havasi</surname>
          </string-name>
          .
          <year>2017</year>
          .
          <article-title>Conceptnet 5.5: An open multilingual graph of general knowledge</article-title>
          .
          <source>In Thirty-first AAAI conference on artificial intelligence .</source>
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          <string-name>
            <surname>Tiiu</surname>
            <given-names>U</given-names>
          </string-name>
          ¨ ksik, Jelena Kallas, Kristina Koppel, Katrin Tsepelina, and
          <string-name>
            <given-names>Raili</given-names>
            <surname>Pool</surname>
          </string-name>
          .
          <year>2021</year>
          .
          <article-title>Estonian as a second language teacher's tools</article-title>
          .
          <source>In Proceedings of the 16th Workshop on Innovative Use of NLP for Building Educational Applications</source>
          , pages
          <fpage>130</fpage>
          -
          <lpage>134</lpage>
          , Online, April. Association for Computational Linguistics.
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          <string-name>
            <given-names>Piek</given-names>
            <surname>Vossen</surname>
          </string-name>
          .
          <year>1996</year>
          .
          <article-title>Right or wrong: Combining lexical resources in the eurowordnet project</article-title>
          .
          <source>In Proceedings of the 7th EURALEX International Congress</source>
          , pages
          <fpage>715</fpage>
          -
          <lpage>728</lpage>
          , aug.
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          <string-name>
            <given-names>Piek</given-names>
            <surname>Vossen</surname>
          </string-name>
          .
          <year>1998</year>
          .
          <article-title>Introduction to eurowordnet. In EuroWordNet: A multilingual database with lexical semantic networks</article-title>
          , pages
          <fpage>1</fpage>
          -
          <lpage>17</lpage>
          . Springer.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>