<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Spanish Word Embeddings Learned on Word Association Norms</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Helena Gomez-Adorno</string-name>
          <email>helena.gomez@iimas.unam.mx</email>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Jorge R</string-name>
          <email>jorge.reyes@correo.uady.mx</email>
        </contrib>
        <contrib contrib-type="author">
          <string-name>l-Enguix</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Facultad de Matematicas, Universidad Autonoma de Yucatan</institution>
          ,
          <addr-line>Merida, Yucatan</addr-line>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Instituto de Ingenier a, Universidad Nacional Autonoma de Mexico</institution>
          ,
          <addr-line>Ciudad de Mexico</addr-line>
          ,
          <country country="MX">Mexico</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Instituto de Investigaciones en Matematicas Aplicadas y en Sistemas, Universidad Nacional Autonoma de Mexico</institution>
          ,
          <addr-line>Ciudad de Mexico</addr-line>
          ,
          <country country="MX">Mexico</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Word embeddings are vector representations of words in an n-dimensional space used for many natural language processing tasks. A large training corpus is needed for learning good quality word embeddings. In this work, we present a method based on the node2vec algorithm for learning embeddings based on paths in a graph. We used a collection of Word Association Norms in Spanish to build a graph of word connections. The nodes of the network correspond to the words in the corpus, whereas the edges correspond to a pair of words given in a free association test. We evaluated our word vectors in human annotated benchmarks, achieving better results than those trained on a billion-word corpus such as, word2vec, fasttext, and glove.</p>
      </abstract>
      <kwd-group>
        <kwd>word vectors</kwd>
        <kwd>node2vec</kwd>
        <kwd>word association norms</kwd>
        <kwd>Spanish</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        The representation of words in a vector space is a very active research area
in the latest decades. Computational models like the singular value
decomposition (SVD) and the latent semantic analysis (LSA) are capable of modeling word
vector representations (word embeddings ) from the term-document matrix. Both
methods can reduce a dataset of N dimensions using only the most important
features. Recently, Mikolov et al. [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ] introduced word2vec inspired by the
distributional hypothesis establishing that words in similar contexts tends to have
similar meanings [
        <xref ref-type="bibr" rid="ref22">22</xref>
        ]. This method uses a neural network in order to learn vector
representations of words by predicting other words in their context. The vector
representation of a word obtained by word2vec has the awesome capability of
conserving linear regularities between words.
      </p>
      <p>In order to build a model of adequate and reliable vector space, capable of
capturing semantic similarity and linear regularities of words, large volumes of
text are needed. Although word2vec is fast and e cient to train, and pre-trained
word vectors are usually available online, it is still computationally expensive
to process large volumes of data in non-commercial environments, that is, on
personal computers.</p>
      <p>
        Free association is an experimental technique commonly used to discover the
way in which the human mind structures knowledge [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. In free association tests,
a person is asked to say the rst word that comes to mind in response to a given
stimulus word. The set of lexical relations obtained with these experiments is
called Word Association Norms (WAN). These kinds of resources re ect both
semantic and episodic contents [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ].
      </p>
      <p>
        In previous work [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] we learn word vectors in English from a graph obtained
from a WAN corpus. The vectors learned from this graph were able to map the
contents of semantic and episodic memory in vector space. For this purpose,
we used the node2vec algorithm [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] which is able to learn node mappings to
a continuous vector space from the complete network taking into account the
neighborhood of the nodes. The algorithm performs biased random paths to
explore di erent neighborhoods in order to capture not only the structural roles
of the nodes in the network but also the communities to which they belong to.
      </p>
      <p>
        In this paper, we extend previous work of learning word vectors in English[
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]
by learning vector representations of words from a resource that collects words
association norms in Spanish. We build two embedding resources of di erent
dimensions, the rst one based on Normas de Asociacion Libre en Castellano [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]
(NALC), and the other using the corpus of Normas de Asociacion de Palabras
para el Espan~ol de Mexico [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] (NAP). The obtained embeddings from both
resources are available on GitHub, the NALC based embeddings4 and the NAP
based embeddings5.
      </p>
      <p>The rest of the paper is organized as follows. In section 2, we discuss the
related work. In Section 3, we present the corpora of Word Association Norms.
In section 4, we describe the methodological framework for learning word vectors
from WAN's. Section 5, shows the evaluation of the generated vectors, using a
word similarity dataset in Spanish. Finally, in section 6 we draw some conclusions
and point out to possible directions of future work.
2</p>
    </sec>
    <sec id="sec-2">
      <title>Related Work</title>
      <p>
        Semantic networks [
        <xref ref-type="bibr" rid="ref25">25</xref>
        ] are graphs relating words [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] used in linguistics and
psycholinguistics not only to study the organization of the vocabulary but also to
approach the structure of knowledge. Many languages have corpora of WAN.
In the past decades, di erent association lists were elaborated with the
collaboration of a large number of volunteers. However, in recent years, the web has
      </p>
      <sec id="sec-2-1">
        <title>4 https://github.com/jocarema/nalc_vectors 5 https://github.com/jocarema/nap_vectors</title>
        <p>
          become a natural way to get data to build such resources. Jeux de Mots6
provides an example in French [
          <xref ref-type="bibr" rid="ref18">18</xref>
          ], whereas the Small World of Words7 contained
datasets in 14 languages at the time of writing.
        </p>
        <p>
          Sinopalnikova and Smrz [
          <xref ref-type="bibr" rid="ref24">24</xref>
          ] showed that WATs are comparable to balanced
text corpora and can replace them in case of absence of a corpus. The authors
presented a methodological framework for building and extending semantic
networks with word association thesaurus (WAT), including a comparison of quality
and information provided by WAT vs. other language resources.
        </p>
        <p>
          Borge-Holthoefer &amp; Arenas [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] used free association information for
extracting semantic similarity relations with a Random Inheritance Model (RIM). The
obtained vectors were compared with LSA-based vector representations and the
WAS (word association space) model. Their results indicate that RIM can
successfully extract word feature vectors from a free association network.
        </p>
        <p>
          In a recent work by De Deyne et al. [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ] the authors introduced a method
for learning word vectors from WANs using a spreading activation approach in
order to encode a semantic structure from the WAN. The authors used part
of the Small World of Words network. The word association-based model was
compared with a word embeddings model (word2vec) using relatedness and
similarity judgments from humans, obtaining an average of 13% of improvement
over the word2vec model.
3
        </p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Word Association Norms in Spanish</title>
      <p>
        Many languages have compilations of word association norms. In the past decades,
some interesting works have been developed with a large number of volunteers.
Among the most well-known English resources accessible on the web are the
Edinburgh Associative Thesaurus8 (EAT) [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ] and the resource of Nelson et
al.9 [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ].
      </p>
      <p>
        For Spanish, there are some corpora of free words association, in this work
we used two WAN resources in Spanish: a) Corpus de Normas de Asociacion
de Palabras para el Espan~ol de Mexico (NAP) [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] and b) Corpus de Normas de
Asociacion Libre en Castellano [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] (NALC).
      </p>
      <p>
        The NAP corpus was elaborated with a group of 578 native Mexican speakers
young adults, 239 men and 339 women, with ages ranging from 18 to 28 years,
and with a range of education of at least 11 years. The total number of tokens
in the corpus is 65731, with 4704 di erent words. The authors used 234 stimulus
words, all of them common nouns taken from the MacArthur word compression
and production [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ]. It is important to mention that although the stimuli are
always nouns, the associated words are free-choice, that is, the informants can
relate to the word stimulus with any word regardless of its grammatical category.
      </p>
      <sec id="sec-3-1">
        <title>6 http://www.jeuxdemots.org/ 7 https://smallworldofwords.org/ 8 http://www.eat.rl.ac.uk/ 9 http://web.usf.edu/FreeAssociation</title>
        <p>For each stimuli and its associates, the authors computed di erent measures:
time, frequency and association strength.</p>
        <p>The NALC corpus includes 5819 stimuli words and their corresponding
associates obtained from the free association responses of a sample of 525 subjects
for 247 words, of 200 subjects for 664 words and of 100 for the remaining words.
In the compilation of association norms, approximately 1500 university students
have participated so far. All the subjects had Spanish as their native language
and participated voluntarily in the empirical study. The total number of di erent
words in the corpus is 31207.
4</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Learning Word Embeddings on Spanish WANs</title>
      <p>The graph that represents a given WAN corpus is formally de ned as G =
fV; E; g where:
{ V = fviji = 1; :::; ng is the nite set of nodes with size n, V 6= ;, which
corresponds to stimuli words along with its associates.
{ E = f(vi; vj )jvi; vj 2 V; 1 i; j ng, is the set of edges, which corresponds
to the connections between stimuli and associates words.
{ : E ! R, is a weighting function over the edges.</p>
      <p>We performed experiments with directed and non-directed graphs. In the
directed graphs, each pair of nodes (vi; vj ) follows an established order where
the initial node vi corresponds to the stimulus word and the nal node vj to an
associated word. For the non-directed graph, all the stimuli are connected with
their correspondent associates without any order of precedence. We evaluated
three edges weighting functions:
Time It measures the seconds the participant takes to give an answer for each
stimulus.</p>
      <p>Frequency It establishes the number of occurrences of each of the associated
words with a stimulus. In this work we use the inverse frequency (IF ):
IF =</p>
      <p>F</p>
      <p>F
where F the frequency of a given associated word, and F is the sum of the
frequencies of the words connected to the same stimulus
Association Strength Establishes a relation between the frequency and the
number of responses for each stimulus. It can be calculated as follows:
ASW =</p>
      <p>AW</p>
      <p>100
F
where AW is the frequency of a given word associated with a stimulus, and</p>
      <p>F the sum of the frequencies of the words connected the same stimulus
(the total number of answers). We also used the inverse of the association
strength (IAS):</p>
      <p>The NAP corpus provides the three weighting functions, however for the
NALC corpus only the association strength is available. Thus, in our evaluation
we only report results using the association strength for the NALC corpus.
4.1</p>
      <p>
        Node2vec
Node2vec [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] nds a mapping f : V ! Rd that transforms the nodes of a graph
into vectors of d-dimensions. It de nes a neighborhood in a network Ns(u) V
for each node u 2 V through a S sampling strategy. The goal of the algorithm
is to maximize the probability of observing subsequent nodes on a random path
of a xed length.
      </p>
      <p>The sampling strategy designed in node2vec allows it to explore
neighborhoods with skewed random paths. The parameters p and q control the change
between the breadth- rst search (BFS) and depth- rst search (DFS) in the graph.
Thus, choosing an adequate balance allows preserving both the structure of
the community and the equivalence between structural nodes in the new vector
space.</p>
      <p>In this work, we used the implementation of the project node2vec, which is
available on the web10 with default values for all parameters. We also examined
the quality of vectors with a di erent number of dimensions.
5</p>
    </sec>
    <sec id="sec-5">
      <title>Spanish Word Embeddings Evaluation</title>
      <p>
        There are several evaluation methods for unsupervised word embeddings
methodologies [
        <xref ref-type="bibr" rid="ref23">23</xref>
        ], which are categorized as extrinsic and intrinsic. In the extrinsic
evaluation, the quality of the word vectors is evaluated by the improvement of
performance in a given natural language processing tasks (PLN) [
        <xref ref-type="bibr" rid="ref12 ref13">12, 13</xref>
        ].
Intrinsic evaluation measures the ability of word vectors to capture syntactic or
semantic relationships [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ].
      </p>
      <p>The hypothesis of the intrinsic evaluation is that similar words should have
similar representations. So, we rst performed a visualization of a sample of
words using the T-SNE projection of the word vectors in a two-dimensional
vector space. Figure 1 shows how the words that are related to each other are
grouped. We show the word vectors obtained from graphs with the three
weighting functions using the NAP corpus only. It is observed that in all cases the
vectors illustrate some interesting phenomena. For example, when frequency is
taken as weight (the graph below), the word pajaro (bird) is drawn very close
to avion (plane). From this, it is inferred that the feature \ y" is more
representative than \animal" for the model. For its part, the word caballo (horse), is
represented closer to camioneta (truck) than to other animals, focusing more on
its status as \transportation".</p>
      <p>
        In addition, we evaluated the ability of word vectors to capture semantic
relationships through a word similarity task. Speci cally, we used two widely
10 http://snap.stanford.edu/node2vec/
known corpora: a) the corpus WordSim-353 [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] composed of pairs of terms
semantically related with similarity scores given by humans and b) the MC-30 [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ]
benchmark containing 30 word pairs. Both datasets in its Spanish version 11 [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ].
      </p>
      <p>We calculated the cosine similarity between the vectors of word pairs
contained in the above mentioned datasets and compare it with the similarity given
by humans using the Spearman correlation. To deal with the non-inclusion of
every word of the testing data sets in our NALC word association norms, we
introduced the concept of overlap in the experiments and calculated the total
number of common words between the lists that are being compared. The
others are excluded from the evaluation. In principle, having large overlaps is a
positive feature this approach. Tables 1 and 2 present the Spearman
corre11 http://web.eecs.umich.edu/~mihalcea/downloads.html
lation, of the similarity given by human taggers, with the similarity obtained
with word vectors (learned from NAP and NALC separately). We also report
di erent dimensions of word vectors learned on the non-directed graphs with
di erent weighting functions. We also report the overlap, which is the number
of words that can be found in in both, the given WAN corpus (NAP or NALC)
and the evaluation dataset (ES-WS-53 or MC-30).</p>
      <p>It can be observed that the word embeddings obtained from the NALC corpus
achieved better correlation with the human similarities than the embeddings
obtained from the NAP corpus in both datasets, ES-WS-53 and MC-30. The
di erence in the results can be explained by the size of the vocabulary in both
WANs, the NALC corpus has higher overlap with both evaluation datasets than
the NAP corpus.</p>
      <p>In order to test and compare the quality of the Spanish word vectors, we
also performed the experiments with pre-trained Spanish vectors12. We selected
three word embeddings models: word2vec13, gloVe14, and fasttext15.</p>
      <p>Table 3 shows the Spearman rank order correlation between the cosine
similarity obtained with word vectors pre-trained in large corpora and the similarity
of humans (obtained from WordSim-353 ) and MC-30 datasets) in comparison
with the correlation between NAP embeddings and the humans rated
similarities. In the same way, Table 4 shows the same comparison with pre-trained word
vectors and the NALC based embeddings.</p>
      <p>
        The highest correlation value was obtained with the vectors trained with the
fasttext [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] model. The vectors trained on the Wikipedia in Spanish obtained
the best results among the pre-trained models. Our method outperformed the
results obtained by the pre-trained vectors when the vectors were learned on the
NALC corpus in both evaluation datasets, ES-WS-353 and MC-30.
We introduced a method for learning Spanish word embeddings from a Corpus of
Word Association Norms. For learning the word vectors, we applied the node2vec
algorithm on the graph of two WAN corpora, NAP and NALC.
      </p>
      <p>We employ weighting functions on the edges of the graph taking into account
three di erent criteria: time, inverse frequency and inverse associative strength.
The best results have been obtained with the association strength, however,
the time weighting function also achieved high results. Words with a higher
12 https://github.com/uchile-nlp/spanish-word-embeddings
13 https://code.google.com/archive/p/word2vec/
14 https://nlp.stanford.edu/projects/glove/
15 https://github.com/facebookresearch/fastText/blob/master/
pretrained-vectors.md
association strength usually have a shorter formulation time, which leads to the
algorithm to connect more related words in a neighborhood because the node2vec
algorithm looks for shorter paths in the graphs.</p>
      <p>
        The results we obtained using the NALC corpus are higher than those
obtained with pre-trained word embeddings trained on large corpora. The
performance even improves the results achieved with the vectors trained on the Spanish
billion words corpus [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. However, some simple strategies would help improve our
results. Some of them would be to adjust the parameters of the algorithm and
adapt the system to di erent types of neighborhoods for the nodes, which could
produce di erent con gurations of the vectors. In future work we will perform an
extrinsic evaluation these Spanish word vectors, i.e. in some Natural Language
Processing task [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ].
      </p>
      <p>The evaluations carried out with the vectors learned on the NAP corpus
also showed promising results with respect to the similarity and relational
indexes. However, due to the low vocabulary length, the results were lower than
those obtained on pre-trained embeddings. As future work, we plan to solve
this problem by automatically generate word association norms between pairs of
words retrieved from a medium-sized corpus. With this process, we will build a
new resource that can account for syntactic, semantic and cognitive connections
between words.</p>
    </sec>
    <sec id="sec-6">
      <title>Acknowledgments</title>
      <p>This work was partially supported by the following projects: Conacyt
FC-201601-2225 and PAPIIT IA401219, IN403016, AG400119.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Aitchison</surname>
            ,
            <given-names>J.:</given-names>
          </string-name>
          <article-title>Words in the mind: An introduction to the mental lexicon</article-title>
          . John Wiley &amp; Sons (
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Arias-Trejo</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Barron-Mart nez</surname>
          </string-name>
          , J.B.,
          <string-name>
            <surname>Alderete</surname>
            ,
            <given-names>R.H.L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Aguirre</surname>
            ,
            <given-names>F.A.R.</given-names>
          </string-name>
          : Corpus de normas de asociacion de palabras para el espaol de Mxico [NAP]. Universidad Nacional Autnoma de Mxico (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Baroni</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dinu</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kruszewski</surname>
          </string-name>
          , G.:
          <article-title>Don't count, predict! a systematic comparison of context-counting vs. context-predicting semantic vectors</article-title>
          . In:
          <article-title>Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)</article-title>
          .
          <source>vol. 1</source>
          , pp.
          <volume>238</volume>
          {
          <issue>247</issue>
          (
          <year>2014</year>
          ), http://www.aclweb.org/anthology/ P14-1023
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Bel-Enguix</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gomez-Adorno</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <article-title>Reyes-Magan~a</article-title>
          , J.,
          <string-name>
            <surname>Sierra</surname>
          </string-name>
          , G.:
          <article-title>Wan2vec: Embeddings learned on word association norms</article-title>
          .
          <source>Semantic Web</source>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Bojanowski</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Grave</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Joulin</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mikolov</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          :
          <article-title>Enriching word vectors with subword information</article-title>
          .
          <source>Computing Research Repository arXiv:1607.04606</source>
          (
          <year>2016</year>
          ). https://doi.org/10.1162/tacl a 00051, https://arxiv.org/abs/1607.04606
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Borge-Holthoefer</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Arenas</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Navigating word association norms to extract semantic information</article-title>
          .
          <source>In: Proceedings of the 31st Annual Conference of the Cognitive Science Society</source>
          (
          <year>2009</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Cardellino</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <source>Spanish Billion Words Corpus and Embeddings (March</source>
          <year>2016</year>
          ), http://crscardellino.github.io/SBWCE/
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>De Deyne</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Navarro</surname>
            ,
            <given-names>D.J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Storms</surname>
          </string-name>
          , G.:
          <article-title>Associative strength and semantic activation in the mental lexicon: Evidence from continued word associations</article-title>
          .
          <source>In: Proceedings of the 35th Annual Conference of the Cognitive Science Society. Cognitive Science Society</source>
          (
          <year>2013</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>De Deyne</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Perfors</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Navarro</surname>
            ,
            <given-names>D.J.:</given-names>
          </string-name>
          <article-title>Predicting human similarity judgments with distributional models: The value of word associations</article-title>
          .
          <source>In: Proceedings of COLING</source>
          <year>2016</year>
          ,
          <source>the 26th International Conference on Computational Linguistics: Technical Papers</source>
          . pp.
          <year>1861</year>
          {
          <year>1870</year>
          (
          <year>2016</year>
          ). https://doi.org/10.24963/ijcai.
          <year>2017</year>
          /671
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Fernandez</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>D ez</surname>
          </string-name>
          , E.,
          <string-name>
            <surname>Alonso</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          : Normas de asociacion libre en castellano de la universidad de salamanca (
          <year>2010</year>
          ), http://inico.usal.es/usuarios/gimc/ normas/index_nal.asp
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Finkelstein</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gabrilovich</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Matias</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rivlin</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Solan</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wolfman</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ruppin</surname>
          </string-name>
          , E.:
          <article-title>Placing search in context: The concept revisited</article-title>
          .
          <source>In: Proceedings of the 10th International Conference on World Wide Web</source>
          . pp.
          <volume>406</volume>
          {
          <fpage>414</fpage>
          .
          <string-name>
            <surname>ACM</surname>
          </string-name>
          (
          <year>2001</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Gomez-Adorno</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Markov</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sidorov</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Posadas-Duran</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sanchez-Perez</surname>
            ,
            <given-names>M.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chanona-Hernandez</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          :
          <article-title>Improving feature representation based on a neural network for author pro ling in social media texts</article-title>
          .
          <source>Computational Intelligence and Neuroscience</source>
          <year>2016</year>
          , 13 pages (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Gomez-Adorno</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Posadas-Duran</surname>
            ,
            <given-names>J.P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sidorov</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pinto</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>Document embeddings learned on various types of n-grams for cross-topic authorship attribution</article-title>
          . Computing pp.
          <volume>1</volume>
          {
          <issue>16</issue>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Grover</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Leskovec</surname>
          </string-name>
          , J.: node2vec:
          <article-title>Scalable feature learning for networks</article-title>
          .
          <source>In: Proceedings of the 22nd ACM International Conference on Knowledge Discovery and Data Mining</source>
          . pp.
          <volume>855</volume>
          {
          <fpage>864</fpage>
          .
          <string-name>
            <surname>ACM</surname>
          </string-name>
          (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Hassan</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mihalcea</surname>
          </string-name>
          , R.:
          <article-title>Cross-lingual semantic relatedness using encyclopedic knowledge</article-title>
          .
          <source>In: Proceedings of the 2009 Conference on Empirical Methods in Natural Language Processing: Volume 3-</source>
          Volume 3. pp.
          <volume>1192</volume>
          {
          <fpage>1201</fpage>
          .
          <article-title>Association for Computational Linguistics (</article-title>
          <year>2009</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <surname>Jackson-Maldonado</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Thal</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fenson</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Marchman</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Newton</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Conboy</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          :
          <article-title>Macarthur inventarios del desarrollo de habilidades comunicativas (inventarios): Users guide and technical manual</article-title>
          . Baltimore, MD:
          <string-name>
            <surname>Brookes</surname>
          </string-name>
          (
          <year>2003</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          17.
          <string-name>
            <surname>Kiss</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Armstrong</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Milroy</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Piper</surname>
            ,
            <given-names>J.:</given-names>
          </string-name>
          <article-title>An associative thesaurus of English and its computer analysis</article-title>
          . Edinburgh University Press, Edinburgh (
          <year>1973</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          18.
          <string-name>
            <surname>Lafourcade</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Making people play for lexical acquisition</article-title>
          .
          <source>In Proceedings of the th SNLP</source>
          <year>2007</year>
          , Pattaya, Thaland
          <volume>7</volume>
          ,
          <issue>13</issue>
          {
          <issue>15</issue>
          (
          <year>December 2007</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          19.
          <string-name>
            <surname>Mikolov</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Corrado</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dean</surname>
            ,
            <given-names>J.:</given-names>
          </string-name>
          <article-title>E cient estimation of word representations in vector space</article-title>
          .
          <source>Computing Research Repository arXiv:1301.3781</source>
          (
          <year>2013</year>
          ), https://arxiv.org/abs/1301.3781
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          20.
          <string-name>
            <surname>Miller</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Charlees</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          :
          <article-title>Contextual correlates of semantic similarity</article-title>
          .
          <source>Language and cognitive processes 6</source>
          (
          <issue>1</issue>
          ),
          <volume>1</volume>
          {
          <fpage>28</fpage>
          (
          <year>1991</year>
          ). https://doi.org/10.1080/01690969108406936
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          21.
          <string-name>
            <surname>Nelson</surname>
            ,
            <given-names>D.L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>McEvoy</surname>
            ,
            <given-names>C.L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Schreiber</surname>
            ,
            <given-names>T.A.</given-names>
          </string-name>
          :
          <article-title>Word association rhyme and word fragment norms</article-title>
          . The University of South Florida (
          <year>1998</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          22.
          <string-name>
            <surname>Sahlgren</surname>
            ,
            <given-names>M.:</given-names>
          </string-name>
          <article-title>The distributional hypothesis</article-title>
          .
          <source>Italian Journal of Disability Studies</source>
          <volume>20</volume>
          ,
          <volume>33</volume>
          {
          <fpage>53</fpage>
          (
          <year>2008</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          23.
          <string-name>
            <surname>Schnabel</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Labutov</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mimno</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Joachims</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          :
          <article-title>Evaluation methods for unsupervised word embeddings</article-title>
          .
          <source>In: Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing</source>
          . pp.
          <volume>298</volume>
          {
          <issue>307</issue>
          (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          24.
          <string-name>
            <surname>Sinopalnikova</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Smrz</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Word association thesaurus as a resource for extending semantic networks</article-title>
          . pp.
          <volume>267</volume>
          {
          <issue>273</issue>
          (
          <year>2004</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          25.
          <string-name>
            <surname>Sowa</surname>
            ,
            <given-names>J.F.</given-names>
          </string-name>
          :
          <article-title>Conceptual graphs as a universal knowledge representation</article-title>
          .
          <source>Computers &amp; Mathematics with Applications</source>
          <volume>23</volume>
          (
          <issue>2</issue>
          ),
          <volume>75</volume>
          {
          <fpage>93</fpage>
          (
          <year>1992</year>
          ). https://doi.org/10.1016/
          <fpage>0898</fpage>
          -
          <lpage>1221</lpage>
          (
          <issue>92</issue>
          )
          <fpage>90137</fpage>
          -
          <lpage>7</lpage>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>