<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>ANDI @ CONcreTEXT: Predicting concreteness in context for English and Italian using distributional models and behavioural norms</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Armand Stefan Rotaru Independent researcher</string-name>
          <email>armand.rotaru@gmail.com</email>
        </contrib>
      </contrib-group>
      <abstract>
        <p>In this paper we describe our participation in the CONcreTEXT task of EVALITA 2020, which involved predicting subjective ratings of concreteness for words presented in context. Our approach, which ranked first in both the English and Italian subtasks, relies on a combination of context-dependent and context-independent distributional models, together with behavioural norms. We show that good results can be obtained for Italian, by first automatically translating the Italian stimuli into English, and then using existing resources for both Italian and English.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        In our everyday life we rarely encounter words in
isolation. Instead, we typically process words as
part of sentences or phrases, and these linguistic
contexts shape our understanding of individual
words. However, for various reasons, the
overwhelming majority of behavioural norms that
have been collected so far focus only on single
words or word pairs
        <xref ref-type="bibr" rid="ref7">(Johns et al., 2020)</xref>
        .
      </p>
      <p>
        Thus, the EVALITA 2020
        <xref ref-type="bibr" rid="ref1">(Basile et al., 2020)</xref>
        CONcreTEXT Task
        <xref ref-type="bibr" rid="ref5">(Gregori et al., 2020)</xref>
        represents a timely and valuable contribution to the
study of context-dependent semantics. The task
asks competitors to predict subjective ratings of
concreteness for words presented within
sentences. As mentioned by the organizers, being
able to automatically compute contextual
concreteness ratings would have a several practical
applications, such as identifying the use of
figurative language, detecting words that might be
difficult to understand for language learners, and
allowing tighter control of contextual variables in
psycholinguistic experiments.
      </p>
      <p>In this paper we describe our computational
models, based on pre-trained distributional
models and behavioural norms, which ranked first in
both the English and Italian tracks of the
competition1. We find that the best performance can be
obtained by employing a combination of
transformer models, developed in the last 2 years.
Moreover, for Italian, it is possible to reach good
levels of performance by relying on both the
original stimuli and their English translation, which
allows access to resources for both languages.
1.1</p>
    </sec>
    <sec id="sec-2">
      <title>General description</title>
      <p>In order to predict concreteness in context, we use
information derived from three type of sources,
namely behavioural norms and distributional
models, both context-independent (i.e., a model
outputs the same vector representation for a given
word, regardless of the context in which the word
is encountered), and context-dependent (i.e., a
model outputs a potentially different
representations for a given word, as a function of the context
in which the word is presented).</p>
      <p>
        Firstly, we employ behavioural norms collected
for a wide variety of psycholinguistic factors. Of
particular interest to us are norms for concreteness
        <xref ref-type="bibr" rid="ref2">(Brysbaert et al., 2014)</xref>
        , semantic diversity
        <xref ref-type="bibr" rid="ref6">(Hoffman et al., 2013)</xref>
        , age of acquisition
        <xref ref-type="bibr" rid="ref9">(Kuperman et
al., 2012)</xref>
        , emotional dimensions
        <xref ref-type="bibr" rid="ref15">(i.e., valence,
arousal, and dominance; Mohammad, 2018)</xref>
        , and
sensorimotor dimensions
        <xref ref-type="bibr" rid="ref13">(i.e., modality strengths
for the tactile, auditory, olfactory, gustatory,
visual, and interoceptive modalities; interaction
strengths for the mouth/throat, hand/arm, foot/leg,
head excluding mouth/throat, and torso effectors;
Lynott et al., 2019)</xref>
        , as well as frequency and
contextual diversity counts
        <xref ref-type="bibr" rid="ref21">(Van Heuven et al., 2014)</xref>
        .
      </p>
      <p>We focus on these specific factors since they are
meaningfully related to word concreteness (see
the previous references).</p>
      <p>
        Secondly, we employ context-independent
distributional models, namely Skip-gram
        <xref ref-type="bibr" rid="ref14">(Mikolov
et al., 2013)</xref>
        , FastText (Bojanowski et al., 2017),
GloVe
        <xref ref-type="bibr" rid="ref17">(Pennington et al., 2014)</xref>
        , and ConceptNet
NumberBatch
        <xref ref-type="bibr" rid="ref20">(Speer et al., 2017)</xref>
        . Such models
have been used in order to accurately predict a
range of psycholinguistic variables, including
concreteness
        <xref ref-type="bibr" rid="ref16">(ρ = .88; Paetzold &amp; Specia, 2016)</xref>
        .
      </p>
      <p>
        Thirdly, we employ context-dependent
distributional models, namely BERT (Devlin et al.,
2018), RoBERTa (Liu et al., 2018), AlBERTo
        <xref ref-type="bibr" rid="ref18">(Polignano et al., 2019)</xref>
        , GPT-2
        <xref ref-type="bibr" rid="ref19">(Radford et al.,
2019)</xref>
        , Bart
        <xref ref-type="bibr" rid="ref11">(Lewis et al., 2019)</xref>
        , and ALBERT
(Lan et al., 2020). Although they have become
extremely popular after achieving human-level
performance in various linguistic tasks (e.g., those in
the GLUE benchmark; Wang et al., 2018), we are
not aware of studies looking at whether such
models can accurately predict (contextualized)
subjective ratings. Nevertheless, since these models
were specifically designed to process rich
contextual information, they could be a valuable tool for
predicting ratings of concreteness in context.
1.2
      </p>
    </sec>
    <sec id="sec-3">
      <title>Predictors for English</title>
      <p>We tested (combinations of) three groups of
predictors. The first group was derived from large
datasets of ratings for concreteness, semantic
diversity, age of acquisition, emotional dimensions,
and sensorimotor dimensions, as well as
frequency and contextual diversity counts based on
the SUBTLEX-UK and BNC corpora (see the
references from the beginning of the previous
section). In order to extend the coverage of the
subjective ratings, we did not directly use them as
predictors of concreteness in context. Instead, we
relied on the Skip-gram, GloVe, and ConceptNet
NumberBatch models, as a means of estimating
the subjective ratings for more than 100,000
words, via linear regression. For the frequency
and contextual diversity counts, we kept the
original values, as they already have very good
coverage. The intersection of the two datasets, which
includes more than 70,000 words, served as the
basis for our predictors of concreteness. More
specifically, for each variable V (e.g., semantic
diversity), we generated four predictors, namely V(w),
V(c), V(w) * V(c), and abs(V(w) - V(c)), where:
• V(w) denotes the value of V
corresponding to the word w (e.g., w = “offend”). If
w is not present in our norms, we set V(w)
•
to the average value of V, computed over
the entire norms;
V(c) denotes the value of V corresponding
to the context c in which the word w is
encountered (e.g., w = “offend”, c = “Do
not insult or ___ anyone .“). Computing
this value involves calculating the
average V(c) = ∑!"#$ "($!) , where V(ci) is the
&amp;
value of V corresponding to the i-th
context word, calculated as described
previously, and N is the number of words that
make up the context.</p>
      <p>These predictors allowed us to include both the
individual contributions of word w and its context
c, as well as certain interactions between w and c.</p>
      <p>
        The second group was derived from Skip-gram,
GloVe, and ConceptNet NumberBatch
embeddings, as well as from the concatenation of the
three types of embeddings. The vocabulary of the
four models is that described in the discussion
above. Given the large number of dimensions
involved (i.e., 300 + 300 + 300 + 900 = 1,800), we
first extracted the top 20 principal components
from each model (although comparable results
can also be obtained by using a larger number of
components). Then, for each variable V (e.g., PC3
from the GloVe model) we generated four
predictors, namely V(w), V(c), V(w) * V(c), and abs(V(w)
- V(c)), following the same procedure as in the
previous discussion. In addition, based on
        <xref ref-type="bibr" rid="ref4">(Frassinelli et al., 2017)</xref>
        , for each distributional model
we added four predictors based on a measure of
neighbourhood density (i.e., the mean cosine
similarity between a vector and its closest 20 vectors),
using the same procedure as described above.
      </p>
      <p>The third group was derived from the BERT,
GPT-2, Bart, and ALBERT models. We used the
standard (base) versions of each model (i.e.,
without task-specific fine-tuning), as described in the
original papers, and obtained from the Hugging
Face repository (https://huggingface.co/models).</p>
      <p>Unlike for the previous two groups, the
predictors consist only of a word’s activations from the
last hidden layer (i.e., for the GPT-2, Bart, and
ALBERT models), or averaged from the last four
hidden layers (i.e., for the BERT model).</p>
      <p>Importantly, for each group of predictors we
generated two sets of variables, based on two
versions of the target words (i.e., the words rated by
the participants). In the first set we used the
uninflected form of the target words, taken from the
TARGET column. In contrast, in the second set of
we used the inflected form of the target words,
taken from the words in the TEXT column located
at the positions specified in the INDEX column.
More details can be found in Table 1.</p>
      <p>For predicting ratings of concreteness in
context, we employed ridge regression, with large
values of the parameter lambda (i.e., strong
regularization), after standardized all the variables.
1.3</p>
    </sec>
    <sec id="sec-4">
      <title>Predictors for Italian</title>
      <p>
        Our approach was similar to that for English, but
with certain significant changes, as follows:
• for the first group of predictors, we began
by automatically translating the Italian
stimuli (i.e., the TARGET and TEXT
columns) into English, using the MarianMT
translation model
        <xref ref-type="bibr" rid="ref8">(Junczys-Dowmunt et
al., 2018)</xref>
        . Next, for the translated stimuli
we derived the predictors using the exact
same procedure as in the case of English;
• for the second group of predictors, we
employed Italian versions of the FastText and
ConceptNet NumberBatch models),
together with their concatenation. We
derived the predictors based on the top 30
principal components for each model,
rather than the top 20 principal components,
as in the case of English (although
comparable results can also be obtained by using
a larger number of components);
• for the third group of predictors, we again
employed the English translations and
relied on the same models as for English, and
also the RoBERTa model. For the BERT
model, we only used the activations from
the last hidden layer. We also added the
AlBERTo model, but with the Italian stimuli.
      </p>
      <p>As in the case for English, we generated two
sets of predictors, using either the uninflected or
inflected forms of the target words, together with
their corresponding English translations. More
details can be found in Table 1.</p>
      <p>Once more, we employed ridge regression,
with large values of the parameter lambda (i.e.,
strong regularization), after standardizing all the
variables.
2</p>
      <sec id="sec-4-1">
        <title>Results and discussion</title>
        <p>The results for English and Italian are shown in
Figures 1 and 2, respectively, for various sets of
predictors and regularization strengths. Results
are averaged over 1,000 rounds of 5-fold
crossvalidation, using only the training dataset.</p>
        <p>For English, the results indicate that
contextdependent models (Fig. 1c-d) outperform
behavioural norms (Fig. 1a) and context-independent
models (Fig. 1b). For the latter, even though we
introduced contextual variables by averaging a
given variable (e.g., concreteness) over the words
that make up the context, it appears that this
simple average does not properly capture contextual
information and/or interactions between single
word and contextual information. The addition the
behavioural norms and/or context-independent
models has a negligible effect on performance
(Fig. 1e). In this respect, the excellent results for
context-dependent models are likely due to
several factors, such as the highly non-linear
integration of contextual information, the use of attention
mechanisms, and that of more sophisticated
learning objectives (e.g., next sentence prediction).</p>
        <p>Interestingly, predictors based on inflected
targets consistently outperform those based on
uninflected targets, especially for the
context-dependent models. This shows that morphological
information can be quite valuable. Also, even for the
largest sets of predictors, consisting of more than
3,200 variables per 80 data points, the degree of
regularization appears to matter very little,
indicating surprisingly small levels of overfitting.</p>
        <p>In the case of Italian, the findings are somewhat
different from those for English. Performance is
roughly 10% lower than that for English. This is
expected, given that perfect translation from
Italian to English is impossible, and that the majority
of predictors depend on this translation. The gaps
in performance between predictors for inflected vs
uninflected targets (Fig. 2c-d), and between the
various classes of predictors (Fig. 2a-e), are also
smaller. Moreover, the performance of
contextdependent models can be increased to a small
degree by adding behavioural norms and/or
contextindependent models (Fig. 2f).</p>
        <p>Our best models, as described in Figures 1 and
2, ranked first in both the English track (ρ = .83),
and the Italian track (ρ = .75). The two
correlations are smaller than those for the best models in
the two figures, but this is likely to be an effect of
distributional differences between the training set
and the test set.
3</p>
      </sec>
      <sec id="sec-4-2">
        <title>Conclusion</title>
        <p>Our results suggest that a variety of approaches
can be quite successfully employed in order to
predict concreteness in context. The most
effective predictors are those derived from
context-dependent models (e.g., BERT), but relatively good
results can be obtained also by using
context-independent models (e.g., Skip-gram) and
behavioural norms (e.g., ratings of semantic diversity).</p>
        <p>Such an approach works very well for English,
but less so for Italian, where the range of available
predictors (i.e., pre-trained distributional models
and large behavioural norms) is limited. One
surprisingly effective solution to this problem is to
simply translate the Italian stimuli into English, by
relying on a neural machine translation system
(e.g., MarianMT), and then make use of existing
predictors for English. As an alternative to
translating stimuli, it would be interesting to test
whether comparable results can be obtained using
multilingual versions of context-dependent
models, such as BERT.</p>
      </sec>
      <sec id="sec-4-3">
        <title>Acknowledgements</title>
        <p>We would like to thank the anonymous reviewers,
for their comments and suggestions, as well as the
organizers of the competition, for their support.
Behavioural norms (frequency, etc.)
FastText (Common Crawl + Wikipedia)
ConceptNet NumberBatch
(ConceptNet + Skip-gram + GloVe)
Concatenation of FastText and
ConceptNet NumberBatch
ALBERT (last hidden layer)
AlBERTo (last hidden layer)
Bart (last hidden layer)
BERT (last hidden layer)
GPT-2 (last hidden layer)
RoBERTa (last hidden layer)</p>
        <p>Wang, A., Singh, A., Michael, J., Hill, F., Levy, O., &amp;
Bowman, S. (2018). GLUE: A multi-task
benchmark and analysis platform for natural language
understanding. In T. Linzen, G. Chrupała, &amp; A.
Alishahi (Eds.), Proceedings of the EMNLP
Workshop BlackboxNLP (pp. 353-355). Stroudsburg, PA:
ACL.</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <string-name>
            <surname>Basile</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Croce</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Di</surname>
            <given-names>Maro</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            ,
            <surname>Passaro</surname>
          </string-name>
          ,
          <string-name>
            <surname>L.C.</surname>
          </string-name>
          ,
          <year>2020</year>
          . EVALITA 2020:
          <article-title>Overview of the 7th Evaluation Campaign of Natural Language Processing and Speech Tools for Italian</article-title>
          , in: Basile,
          <string-name>
            <given-names>V.</given-names>
            ,
            <surname>Croce</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            ,
            <surname>Di</surname>
          </string-name>
          <string-name>
            <surname>Maro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            ,
            <surname>Passaro</surname>
          </string-name>
          , L.C. (Eds.),
          <source>Proceedings of 7th Evaluation Campaign of Natural Language Processing and Speech Tools for Italian. Final Workshop (EVALITA</source>
          <year>2020</year>
          ). CEUR.org, Online.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          <string-name>
            <surname>Brysbaert</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Warriner</surname>
            ,
            <given-names>A. B.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Kuperman</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          (
          <year>2014</year>
          ).
          <article-title>Concreteness ratings for 40 thousand generally known English word lemmas</article-title>
          .
          <source>Behavior Research Methods</source>
          ,
          <volume>46</volume>
          (
          <issue>3</issue>
          ),
          <fpage>904</fpage>
          -
          <lpage>911</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          <string-name>
            <surname>Devlin</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chang</surname>
            ,
            <given-names>M. W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lee</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Toutanova</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          (
          <year>2019</year>
          ).
          <article-title>BERT: Pre-training of deep bidirectional transformers for language understanding</article-title>
          . In J. Burstein,
          <string-name>
            <given-names>C.</given-names>
            <surname>Doran</surname>
          </string-name>
          , &amp; T. Solorio (Eds.),
          <source>Proceedings of the NAACL-HLT</source>
          (pp.
          <fpage>4171</fpage>
          -
          <lpage>4186</lpage>
          ). Stroudsburg, PA: ACL.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          <string-name>
            <surname>Frassinelli</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Naumann</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Utt</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          , &amp; im
          <string-name>
            <surname>Walde</surname>
            ,
            <given-names>S. S.</given-names>
          </string-name>
          (
          <year>2017</year>
          ).
          <article-title>Contextual characteristics of concrete and abstract words</article-title>
          . In C. Gardent &amp; C. Retoré (Eds.),
          <source>Proceedings of the IWCS</source>
          (pp.
          <fpage>1</fpage>
          -
          <lpage>7</lpage>
          ). Stroudsburg, PA: ACL.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          <string-name>
            <surname>Gregori</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Montefinese</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Radicioni</surname>
            ,
            <given-names>D. P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ravelli</surname>
            ,
            <given-names>A. A.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Varvara</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          (
          <year>2020</year>
          ).
          <article-title>CONcreTEXT @ Evalita2020: the Concreteness in Context Task</article-title>
          . In V. Basile,
          <string-name>
            <given-names>D.</given-names>
            <surname>Croce</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. Di</given-names>
            <surname>Maro</surname>
          </string-name>
          , &amp; L. C. Passaro (Eds.),
          <source>Proceedings of the 7th Evaluation Campaign of Natural Language Processing and Speech tools for Italian (EVALITA</source>
          <year>2020</year>
          ). Online: CEUR.org.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          <string-name>
            <surname>Hoffman</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lambon</surname>
            <given-names>Ralph</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>M. A.</given-names>
            , &amp;
            <surname>Rogers</surname>
          </string-name>
          ,
          <string-name>
            <surname>T. T.</surname>
          </string-name>
          (
          <year>2013</year>
          ).
          <article-title>Semantic diversity: A measure of semantic ambiguity based on variability in the contextual usage of words</article-title>
          .
          <source>Behavior Research Methods</source>
          ,
          <volume>45</volume>
          (
          <issue>3</issue>
          ),
          <fpage>718</fpage>
          -
          <lpage>730</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          <string-name>
            <surname>Johns</surname>
            ,
            <given-names>B. T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jamieson</surname>
            ,
            <given-names>R. K.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Jones</surname>
            ,
            <given-names>M. N.</given-names>
          </string-name>
          (
          <year>2020</year>
          ).
          <article-title>The continued importance of theory: Lessons from big data approaches to language and cognition</article-title>
          . In S. E. Woo,
          <string-name>
            <given-names>R.</given-names>
            <surname>Proctor</surname>
          </string-name>
          , &amp; L.
          <string-name>
            <surname>Tay</surname>
          </string-name>
          (Eds.),
          <article-title>Big data methods for psychological research: New horizons and challenges</article-title>
          (pp.
          <fpage>277</fpage>
          -
          <lpage>295</lpage>
          ). Washington, DC: APA.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          <string-name>
            <surname>Junczys-Dowmunt</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Grundkiewicz</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dwojak</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hoang</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Heafield</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Neckermann</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Seide</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Germann</surname>
            ,
            <given-names>U.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Aji</surname>
            ,
            <given-names>A. F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bogoychev</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Martins</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Birch</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          (
          <year>2018</year>
          ).
          <article-title>Marian: Fast neural machine translation in C++</article-title>
          . In F. Liu &amp; T. Solorio (Eds.),
          <source>Proceedings of the ACL - System Demonstrations</source>
          (pp.
          <fpage>116</fpage>
          -
          <lpage>121</lpage>
          ). Stroudsburg, PA: ACL.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          <string-name>
            <surname>Kuperman</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Stadthagen-Gonzalez</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Brysbaert</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          (
          <year>2012</year>
          ).
          <article-title>Age-of-acquisition ratings for 30,000 English words</article-title>
          .
          <source>Behavior Research Methods</source>
          ,
          <volume>44</volume>
          (
          <issue>4</issue>
          ),
          <fpage>978</fpage>
          -
          <lpage>990</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          <string-name>
            <surname>Lan</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Goodman</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gimpel</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sharma</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Soricut</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          (
          <year>2019</year>
          ).
          <article-title>ALBERT: A lite BERT for self-supervised learning of language representations</article-title>
          .
          <source>In Proceedings of the ICLR</source>
          (pp.
          <fpage>1</fpage>
          -
          <lpage>17</lpage>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          <string-name>
            <surname>Lewis</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Liu</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Goyal</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ghazvininejad</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mohamed</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Levy</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Stoyanov</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Zettlemoyer</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          (
          <year>2019</year>
          ).
          <article-title>Bart: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension</article-title>
          . In D. Jurafsky,
          <string-name>
            <given-names>J.</given-names>
            <surname>Chai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Schluter</surname>
          </string-name>
          , &amp; J.
          <string-name>
            <surname>Tetreault</surname>
          </string-name>
          (Eds.),
          <source>Proceedings of the ACL</source>
          (pp.
          <fpage>7871</fpage>
          -
          <lpage>7880</lpage>
          ). Stroudsburg, PA: ACL.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          <string-name>
            <surname>Liu</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ott</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Goyal</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Du</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Joshi</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Levy</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lewis</surname>
          </string-name>
          . M.,
          <string-name>
            <surname>Zettlemoyer</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Stoyanov</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          (
          <year>2019</year>
          ).
          <article-title>RoBERTa: A robustly optimized BERT pretraining approach</article-title>
          . arXiv preprint:
          <year>1907</year>
          .11692.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          <string-name>
            <surname>Lynott</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Connell</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Brysbaert</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Brand</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Carney</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          (
          <year>2019</year>
          ).
          <article-title>The Lancaster Sensorimotor Norms: Multidimensional measures of perceptual and action strength for 40,000 English words</article-title>
          .
          <source>Behavior Research Methods</source>
          ,
          <volume>52</volume>
          ,
          <fpage>1</fpage>
          -
          <lpage>21</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          <string-name>
            <surname>Mikolov</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Corrado</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Dean</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          (
          <year>2013</year>
          ).
          <article-title>Efficient estimation of word representations in vector space</article-title>
          . In J. Bengio &amp; Y. LeCun (Eds.),
          <source>Proceedings of the Workshop at the ICLR</source>
          (pp.
          <fpage>1</fpage>
          -
          <lpage>12</lpage>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          <string-name>
            <surname>Mohammad</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          (
          <year>2018</year>
          ).
          <article-title>Obtaining reliable human ratings of valence, arousal, and dominance for 20,000 English words</article-title>
          . In I. Gurevych &amp; Y. Miyao (Eds.),
          <source>Proceedings of the ACL - Long Papers</source>
          (pp.
          <fpage>174</fpage>
          -
          <lpage>184</lpage>
          ). Stroudsburg, PA: ACL.
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          <string-name>
            <surname>Paetzold</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Specia</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          (
          <year>2016</year>
          ).
          <article-title>Inferring psycholinguistic properties of words</article-title>
          . In K. Knight,
          <string-name>
            <given-names>A.</given-names>
            <surname>Nenkova</surname>
          </string-name>
          , &amp; O.
          <string-name>
            <surname>Rambow</surname>
          </string-name>
          (Eds.),
          <source>Proceedings of the NAACL-HLT</source>
          (pp.
          <fpage>435</fpage>
          -
          <lpage>440</lpage>
          ). Stroudsburg, PA: ACL.
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          <string-name>
            <surname>Pennington</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Socher</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Manning</surname>
            ,
            <given-names>C. D.</given-names>
          </string-name>
          (
          <year>2014</year>
          ).
          <article-title>GloVe: Global vectors for word representation</article-title>
          . In A. Moschitti,
          <string-name>
            <given-names>B.</given-names>
            <surname>Pang</surname>
          </string-name>
          , &amp; W. Daelemans (Eds.),
          <source>Proceedings of the EMNLP</source>
          (pp.
          <fpage>1532</fpage>
          -
          <lpage>1543</lpage>
          ). Stroudsburg, PA: ACL.
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          <string-name>
            <surname>Polignano</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Basile</surname>
            , P., de Gemmis,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Semeraro</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Basile</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          (
          <year>2019</year>
          ).
          <article-title>AlBERTo: Italian BERT language understanding model for NLP challenging tasks based on tweets</article-title>
          . In R. Bernardi,
          <string-name>
            <given-names>R.</given-names>
            <surname>Navigli</surname>
          </string-name>
          , &amp; G. Semeraro (Eds.), Proceedings of CLiC-it. Aachen, Germany: CEUR.
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          <string-name>
            <surname>Radford</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wu</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Child</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Luan</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Amodei</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Sutskever</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          (
          <year>2019</year>
          ).
          <article-title>Language models are unsupervised multitask learners</article-title>
          .
          <source>OpenAI blog.</source>
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          <string-name>
            <surname>Speer</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chin</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Havasi</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          (
          <year>2017</year>
          ).
          <article-title>ConceptNet 5.5: An open multilingual graph of general knowledge</article-title>
          . In S. P. Singh &amp; S. Markovitch (Eds.),
          <source>Proceedings of the AAAI</source>
          (pp.
          <fpage>4444</fpage>
          -
          <lpage>4451</lpage>
          ). Palo Alto, CA: AAAI Press.
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          <string-name>
            <surname>Van Heuven</surname>
            ,
            <given-names>W. J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mandera</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Keuleers</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Brysbaert</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          (
          <year>2014</year>
          ).
          <article-title>SUBTLEX-UK: A new and improved word frequency database for British English</article-title>
          .
          <source>The Quarterly Journal of Experimental Psychology</source>
          ,
          <volume>67</volume>
          (
          <issue>6</issue>
          ),
          <fpage>1176</fpage>
          -
          <lpage>1190</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>