<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Spanish Word Embeddings Learned on Word Association Norms</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Helena</forename><surname>Gómez-Adorno</surname></persName>
							<email>helena.gomez@iimas.unam.mx</email>
							<affiliation key="aff0">
								<orgName type="department">Instituto de Investigaciones en Matemáticas Aplicadas y en Sistemas</orgName>
								<orgName type="institution">Universidad Nacional Autónoma de México</orgName>
								<address>
									<settlement>Ciudad de México</settlement>
									<country key="MX">México</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Jorge</forename><surname>Reyes-Magaña</surname></persName>
							<affiliation key="aff1">
								<orgName type="department">Instituto de Ingeniería</orgName>
								<orgName type="institution">Universidad Nacional Autónoma de México</orgName>
								<address>
									<settlement>Ciudad de México</settlement>
									<country key="MX">México</country>
								</address>
							</affiliation>
							<affiliation key="aff2">
								<orgName type="department">Facultad de Matemáticas</orgName>
								<orgName type="institution">Universidad Autónoma de Yucatán</orgName>
								<address>
									<settlement>Mérida</settlement>
									<region>Yucatán</region>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Gemma</forename><surname>Bel-Enguix</surname></persName>
							<affiliation key="aff1">
								<orgName type="department">Instituto de Ingeniería</orgName>
								<orgName type="institution">Universidad Nacional Autónoma de México</orgName>
								<address>
									<settlement>Ciudad de México</settlement>
									<country key="MX">México</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Gerardo</forename><surname>Sierra</surname></persName>
							<email>gsierram@iingen.unam.mx</email>
							<affiliation key="aff1">
								<orgName type="department">Instituto de Ingeniería</orgName>
								<orgName type="institution">Universidad Nacional Autónoma de México</orgName>
								<address>
									<settlement>Ciudad de México</settlement>
									<country key="MX">México</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Spanish Word Embeddings Learned on Word Association Norms</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">B0BAAD462DF2C7D157E44D943ADF5014</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-25T02:45+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>word vectors</term>
					<term>node2vec</term>
					<term>word association norms</term>
					<term>Spanish</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Word embeddings are vector representations of words in an n-dimensional space used for many natural language processing tasks. A large training corpus is needed for learning good quality word embeddings. In this work, we present a method based on the node2vec algorithm for learning embeddings based on paths in a graph. We used a collection of Word Association Norms in Spanish to build a graph of word connections. The nodes of the network correspond to the words in the corpus, whereas the edges correspond to a pair of words given in a free association test. We evaluated our word vectors in human annotated benchmarks, achieving better results than those trained on a billion-word corpus such as, word2vec, fasttext, and glove.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>The representation of words in a vector space is a very active research area in the latest decades. Computational models like the singular value decomposition (SVD) and the latent semantic analysis (LSA) are capable of modeling word vector representations (word embeddings) from the term-document matrix. Both methods can reduce a dataset of N dimensions using only the most important features. Recently, Mikolov et al. <ref type="bibr" target="#b18">[19]</ref> introduced word2vec inspired by the distributional hypothesis establishing that words in similar contexts tends to have similar meanings <ref type="bibr" target="#b21">[22]</ref>. This method uses a neural network in order to learn vector representations of words by predicting other words in their context. The vector representation of a word obtained by word2vec has the awesome capability of conserving linear regularities between words.</p><p>In order to build a model of adequate and reliable vector space, capable of capturing semantic similarity and linear regularities of words, large volumes of text are needed. Although word2vec is fast and efficient to train, and pre-trained word vectors are usually available online, it is still computationally expensive to process large volumes of data in non-commercial environments, that is, on personal computers.</p><p>Free association is an experimental technique commonly used to discover the way in which the human mind structures knowledge <ref type="bibr" target="#b7">[8]</ref>. In free association tests, a person is asked to say the first word that comes to mind in response to a given stimulus word. The set of lexical relations obtained with these experiments is called Word Association Norms (WAN). These kinds of resources reflect both semantic and episodic contents <ref type="bibr" target="#b5">[6]</ref>.</p><p>In previous work <ref type="bibr" target="#b3">[4]</ref> we learn word vectors in English from a graph obtained from a WAN corpus. The vectors learned from this graph were able to map the contents of semantic and episodic memory in vector space. For this purpose, we used the node2vec algorithm <ref type="bibr" target="#b13">[14]</ref> which is able to learn node mappings to a continuous vector space from the complete network taking into account the neighborhood of the nodes. The algorithm performs biased random paths to explore different neighborhoods in order to capture not only the structural roles of the nodes in the network but also the communities to which they belong to.</p><p>In this paper, we extend previous work of learning word vectors in English[4] by learning vector representations of words from a resource that collects words association norms in Spanish. We build two embedding resources of different dimensions, the first one based on Normas de Asociación Libre en Castellano <ref type="bibr" target="#b9">[10]</ref> (NALC), and the other using the corpus of Normas de Asociación de Palabras para el Español de México <ref type="bibr" target="#b1">[2]</ref> (NAP). The obtained embeddings from both resources are available on GitHub, the NALC based embeddings <ref type="foot" target="#foot_0">4</ref> and the NAP based embeddings <ref type="foot" target="#foot_1">5</ref> .</p><p>The rest of the paper is organized as follows. In section 2, we discuss the related work. In Section 3, we present the corpora of Word Association Norms. In section 4, we describe the methodological framework for learning word vectors from WAN's. Section 5, shows the evaluation of the generated vectors, using a word similarity dataset in Spanish. Finally, in section 6 we draw some conclusions and point out to possible directions of future work.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Related Work</head><p>Semantic networks <ref type="bibr" target="#b24">[25]</ref> are graphs relating words <ref type="bibr" target="#b0">[1]</ref> used in linguistics and psycholinguistics not only to study the organization of the vocabulary but also to approach the structure of knowledge. Many languages have corpora of WAN. In the past decades, different association lists were elaborated with the collaboration of a large number of volunteers. However, in recent years, the web has become a natural way to get data to build such resources. Jeux de Mots<ref type="foot" target="#foot_2">6</ref> provides an example in French <ref type="bibr" target="#b17">[18]</ref>, whereas the Small World of Words<ref type="foot" target="#foot_3">7</ref> contained datasets in 14 languages at the time of writing.</p><p>Sinopalnikova and Smrz <ref type="bibr" target="#b23">[24]</ref> showed that WATs are comparable to balanced text corpora and can replace them in case of absence of a corpus. The authors presented a methodological framework for building and extending semantic networks with word association thesaurus (WAT), including a comparison of quality and information provided by WAT vs. other language resources.</p><p>Borge-Holthoefer &amp; Arenas <ref type="bibr" target="#b5">[6]</ref> used free association information for extracting semantic similarity relations with a Random Inheritance Model (RIM). The obtained vectors were compared with LSA-based vector representations and the WAS (word association space) model. Their results indicate that RIM can successfully extract word feature vectors from a free association network.</p><p>In a recent work by De Deyne et al. <ref type="bibr" target="#b8">[9]</ref> the authors introduced a method for learning word vectors from WANs using a spreading activation approach in order to encode a semantic structure from the WAN. The authors used part of the Small World of Words network. The word association-based model was compared with a word embeddings model (word2vec) using relatedness and similarity judgments from humans, obtaining an average of 13% of improvement over the word2vec model.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Word Association Norms in Spanish</head><p>Many languages have compilations of word association norms. In the past decades, some interesting works have been developed with a large number of volunteers. Among the most well-known English resources accessible on the web are the Edinburgh Associative Thesaurus<ref type="foot" target="#foot_4">8</ref> (EAT) <ref type="bibr" target="#b16">[17]</ref> and the resource of Nelson et al.<ref type="foot" target="#foot_5">9</ref>  <ref type="bibr" target="#b20">[21]</ref>.</p><p>For Spanish, there are some corpora of free words association, in this work we used two WAN resources in Spanish: a) Corpus de Normas de Asociación de Palabras para el Español de México (NAP) <ref type="bibr" target="#b1">[2]</ref> and b) Corpus de Normas de Asociación Libre en Castellano <ref type="bibr" target="#b9">[10]</ref> (NALC).</p><p>The NAP corpus was elaborated with a group of 578 native Mexican speakers young adults, 239 men and 339 women, with ages ranging from 18 to 28 years, and with a range of education of at least 11 years. The total number of tokens in the corpus is 65731, with 4704 different words. The authors used 234 stimulus words, all of them common nouns taken from the MacArthur word compression and production <ref type="bibr" target="#b15">[16]</ref>. It is important to mention that although the stimuli are always nouns, the associated words are free-choice, that is, the informants can relate to the word stimulus with any word regardless of its grammatical category.</p><p>For each stimuli and its associates, the authors computed different measures: time, frequency and association strength.</p><p>The NALC corpus includes 5819 stimuli words and their corresponding associates obtained from the free association responses of a sample of 525 subjects for 247 words, of 200 subjects for 664 words and of 100 for the remaining words. In the compilation of association norms, approximately 1500 university students have participated so far. All the subjects had Spanish as their native language and participated voluntarily in the empirical study. The total number of different words in the corpus is 31207.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Learning Word Embeddings on Spanish WANs</head><p>The graph that represents a given WAN corpus is formally defined as G = {V, E, φ} where:</p><formula xml:id="formula_0">-V = {v i |i = 1, .</formula><p>.., n} is the finite set of nodes with size n, V = ∅, which corresponds to stimuli words along with its associates.</p><formula xml:id="formula_1">-E = {(v i , v j )|v i , v j ∈ V, 1 ≤ i, j ≤ n},</formula><p>is the set of edges, which corresponds to the connections between stimuli and associates words. φ : E → R, is a weighting function over the edges.</p><p>We performed experiments with directed and non-directed graphs. In the directed graphs, each pair of nodes (v i , v j ) follows an established order where the initial node v i corresponds to the stimulus word and the final node v j to an associated word. For the non-directed graph, all the stimuli are connected with their correspondent associates without any order of precedence. We evaluated three edges weighting functions: Time It measures the seconds the participant takes to give an answer for each stimulus. Frequency It establishes the number of occurrences of each of the associated words with a stimulus. In this work we use the inverse frequency (IF ):</p><formula xml:id="formula_2">IF = ΣF − F</formula><p>where F the frequency of a given associated word, and ΣF is the sum of the frequencies of the words connected to the same stimulus Association Strength Establishes a relation between the frequency and the number of responses for each stimulus. It can be calculated as follows:</p><p>AS W = AW * 100 ΣF where AW is the frequency of a given word associated with a stimulus, and ΣF the sum of the frequencies of the words connected the same stimulus (the total number of answers). We also used the inverse of the association strength (IAS):</p><formula xml:id="formula_3">IAS = 1 − F ΣF</formula><p>The NAP corpus provides the three weighting functions, however for the NALC corpus only the association strength is available. Thus, in our evaluation we only report results using the association strength for the NALC corpus.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1">Node2vec</head><p>Node2vec <ref type="bibr" target="#b13">[14]</ref> finds a mapping f : V → R d that transforms the nodes of a graph into vectors of d-dimensions. It defines a neighborhood in a network N s (u) ⊂ V for each node u ∈ V through a S sampling strategy. The goal of the algorithm is to maximize the probability of observing subsequent nodes on a random path of a fixed length.</p><p>The sampling strategy designed in node2vec allows it to explore neighborhoods with skewed random paths. The parameters p and q control the change between the breadth-first search (BFS) and depth-first search (DFS) in the graph. Thus, choosing an adequate balance allows preserving both the structure of the community and the equivalence between structural nodes in the new vector space.</p><p>In this work, we used the implementation of the project node2vec, which is available on the web<ref type="foot" target="#foot_6">10</ref> with default values for all parameters. We also examined the quality of vectors with a different number of dimensions.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Spanish Word Embeddings Evaluation</head><p>There are several evaluation methods for unsupervised word embeddings methodologies <ref type="bibr" target="#b22">[23]</ref>, which are categorized as extrinsic and intrinsic. In the extrinsic evaluation, the quality of the word vectors is evaluated by the improvement of performance in a given natural language processing tasks (PLN) <ref type="bibr" target="#b11">[12,</ref><ref type="bibr" target="#b12">13]</ref>. Intrinsic evaluation measures the ability of word vectors to capture syntactic or semantic relationships <ref type="bibr" target="#b2">[3]</ref>.</p><p>The hypothesis of the intrinsic evaluation is that similar words should have similar representations. So, we first performed a visualization of a sample of words using the T-SNE projection of the word vectors in a two-dimensional vector space. Figure <ref type="figure" target="#fig_0">1</ref> shows how the words that are related to each other are grouped. We show the word vectors obtained from graphs with the three weighting functions using the NAP corpus only. It is observed that in all cases the vectors illustrate some interesting phenomena. For example, when frequency is taken as weight (the graph below), the word pájaro (bird) is drawn very close to avión (plane). From this, it is inferred that the feature "fly" is more representative than "animal" for the model. For its part, the word caballo (horse), is represented closer to camioneta (truck) than to other animals, focusing more on its status as "transportation".</p><p>In addition, we evaluated the ability of word vectors to capture semantic relationships through a word similarity task. Specifically, we used two widely known corpora: a) the corpus WordSim-353 <ref type="bibr" target="#b10">[11]</ref> composed of pairs of terms semantically related with similarity scores given by humans and b) the MC-30 <ref type="bibr" target="#b19">[20]</ref> benchmark containing 30 word pairs. Both datasets in its Spanish version<ref type="foot" target="#foot_7">11</ref>  <ref type="bibr" target="#b14">[15]</ref>.</p><p>We calculated the cosine similarity between the vectors of word pairs contained in the above mentioned datasets and compare it with the similarity given by humans using the Spearman correlation. To deal with the non-inclusion of every word of the testing data sets in our NALC word association norms, we introduced the concept of overlap in the experiments and calculated the total number of common words between the lists that are being compared. The others are excluded from the evaluation. In principle, having large overlaps is a positive feature this approach. Tables <ref type="table" target="#tab_1">1 and 2</ref> present the Spearman corre-lation, of the similarity given by human taggers, with the similarity obtained with word vectors (learned from NAP and NALC separately). We also report different dimensions of word vectors learned on the non-directed graphs with different weighting functions. We also report the overlap, which is the number of words that can be found in in both, the given WAN corpus (NAP or NALC) and the evaluation dataset (ES-WS-53 or MC-30). It can be observed that the word embeddings obtained from the NALC corpus achieved better correlation with the human similarities than the embeddings obtained from the NAP corpus in both datasets, ES-WS-53 and MC-30. The difference in the results can be explained by the size of the vocabulary in both WANs, the NALC corpus has higher overlap with both evaluation datasets than the NAP corpus.</p><p>In order to test and compare the quality of the Spanish word vectors, we also performed the experiments with pre-trained Spanish vectors<ref type="foot" target="#foot_8">12</ref> . We selected three word embeddings models: word2vec<ref type="foot" target="#foot_9">13</ref> , gloVe <ref type="foot" target="#foot_10">14</ref> , and fasttext <ref type="foot" target="#foot_11">15</ref> .</p><p>Table <ref type="table" target="#tab_2">3</ref> shows the Spearman rank order correlation between the cosine similarity obtained with word vectors pre-trained in large corpora and the similarity of humans (obtained from WordSim-353 ) and MC-30 datasets) in comparison with the correlation between NAP embeddings and the humans rated similarities. In the same way, Table <ref type="table" target="#tab_3">4</ref> shows the same comparison with pre-trained word vectors and the NALC based embeddings.</p><p>The highest correlation value was obtained with the vectors trained with the fasttext <ref type="bibr" target="#b4">[5]</ref> model. The vectors trained on the Wikipedia in Spanish obtained the best results among the pre-trained models. Our method outperformed the results obtained by the pre-trained vectors when the vectors were learned on the NALC corpus in both evaluation datasets, ES-WS-353 and MC-30. We introduced a method for learning Spanish word embeddings from a Corpus of Word Association Norms. For learning the word vectors, we applied the node2vec algorithm on the graph of two WAN corpora, NAP and NALC. We employ weighting functions on the edges of the graph taking into account three different criteria: time, inverse frequency and inverse associative strength. The best results have been obtained with the association strength, however, the time weighting function also achieved high results. Words with a higher association strength usually have a shorter formulation time, which leads to the algorithm to connect more related words in a neighborhood because the node2vec algorithm looks for shorter paths in the graphs.</p><p>The results we obtained using the NALC corpus are higher than those obtained with pre-trained word embeddings trained on large corpora. The performance even improves the results achieved with the vectors trained on the Spanish billion words corpus <ref type="bibr" target="#b6">[7]</ref>. However, some simple strategies would help improve our results. Some of them would be to adjust the parameters of the algorithm and adapt the system to different types of neighborhoods for the nodes, which could produce different configurations of the vectors. In future work we will perform an extrinsic evaluation these Spanish word vectors, i.e. in some Natural Language Processing task <ref type="bibr" target="#b3">[4]</ref>.</p><p>The evaluations carried out with the vectors learned on the NAP corpus also showed promising results with respect to the similarity and relational indexes. However, due to the low vocabulary length, the results were lower than those obtained on pre-trained embeddings. As future work, we plan to solve this problem by automatically generate word association norms between pairs of words retrieved from a medium-sized corpus. With this process, we will build a new resource that can account for syntactic, semantic and cognitive connections between words.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Fig. 1 .</head><label>1</label><figDesc>Fig. 1. Projection of the word vectors in 5 semantic groups (of ten words each).</figDesc><graphic coords="6,134.77,115.83,345.83,344.87" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1 .</head><label>1</label><figDesc>Spearman rank order correlations between Spanish WAN embeddings (based on cosine similarity) and the ES-WS-353 dataset.</figDesc><table><row><cell></cell><cell></cell><cell>NAP</cell><cell></cell><cell>NALC</cell></row><row><cell></cell><cell></cell><cell>Overlap 140</cell><cell></cell><cell>Overlap 322</cell></row><row><cell>Dimension</cell><cell>Inv. Frequency</cell><cell>Inv. Association</cell><cell cols="2">Time Inv. Association</cell></row><row><cell>300</cell><cell>0.489</cell><cell>0.463</cell><cell>0.461</cell><cell>0.650</cell></row><row><cell>200</cell><cell>0.454</cell><cell>0.456</cell><cell>0.491</cell><cell>0.641</cell></row><row><cell>128</cell><cell>0.503</cell><cell>0.463</cell><cell>0.450</cell><cell>0.659</cell></row><row><cell>100</cell><cell>0.471</cell><cell>0.478</cell><cell>0.495</cell><cell>0.664</cell></row><row><cell>50</cell><cell>0.523</cell><cell>0.503</cell><cell>0.503</cell><cell>0.626</cell></row><row><cell>25</cell><cell>0.484</cell><cell>0.478</cell><cell>0.572</cell><cell>0.611</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2 .</head><label>2</label><figDesc>Spearman rank order correlations between Spanish WAN embeddings (based on cosine similarity) and the MC-30 dataset</figDesc><table><row><cell></cell><cell></cell><cell>NAP</cell><cell></cell><cell>NALC</cell></row><row><cell></cell><cell></cell><cell>Overlap 11</cell><cell></cell><cell>Overlap 27</cell></row><row><cell>Dimension</cell><cell>Inv. Frequency</cell><cell>Inv. Association</cell><cell cols="2">Time Inv. Asociation</cell></row><row><cell>300</cell><cell>0.305</cell><cell>0.563</cell><cell>0.545</cell><cell>0.837</cell></row><row><cell>200</cell><cell>0.468</cell><cell>0.381</cell><cell>0.263</cell><cell>0.844</cell></row><row><cell>128</cell><cell>0.545</cell><cell>0.272</cell><cell>0.300</cell><cell>0.767</cell></row><row><cell>100</cell><cell>0.336</cell><cell>0.418</cell><cell>0.372</cell><cell>0.806</cell></row><row><cell>50</cell><cell>0.527</cell><cell>0.509</cell><cell>0.272</cell><cell>0.814</cell></row><row><cell>25</cell><cell>0.454</cell><cell>0.400</cell><cell>0.563</cell><cell>0.788</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 3 .</head><label>3</label><figDesc>Spearman rank order correlation comparison of NAP embeddings and pretrained word vectors with the evaluation datasets.</figDesc><table><row><cell>Source</cell><cell>Vector size</cell><cell cols="2">MC-30 (Overlap 11) ES-WS-353 (Overlap 140)</cell></row><row><cell>Fasttext-sbwc</cell><cell>300</cell><cell>0.881</cell><cell>0.639</cell></row><row><cell>Fasttext-wiki</cell><cell>300</cell><cell>0.936</cell><cell>0.701</cell></row><row><cell>Glove-sbwc</cell><cell>300</cell><cell>0.827</cell><cell>0.532</cell></row><row><cell>Word2vec-sbwc</cell><cell>300</cell><cell>0.890</cell><cell>0.634</cell></row><row><cell>n2v-Inverse Association</cell><cell>300</cell><cell>0.563</cell><cell>0.463</cell></row><row><cell>n2v-Inverse Frequency</cell><cell>300</cell><cell>0.305</cell><cell>0.489</cell></row><row><cell>n2v-Time</cell><cell>25</cell><cell>0.563</cell><cell>0.572</cell></row><row><cell cols="3">6 Conclusions and Future Work</cell><cell></cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 4 .</head><label>4</label><figDesc>Spearman rank order correlation comparison of NALC embeddings and pretrained word vectors with the evaluation datasets.</figDesc><table><row><cell>Source</cell><cell>Vector size</cell><cell cols="2">MC-30 (Overlap 27) ES-WS-353 (Overlap 322)</cell></row><row><cell>Fasttext-sbwc</cell><cell>300</cell><cell>0.762</cell><cell>0.613</cell></row><row><cell>Fasttext-wiki</cell><cell>300</cell><cell>0.793</cell><cell>0.624</cell></row><row><cell>Glove-sbwc</cell><cell>300</cell><cell>0.707</cell><cell>0.482</cell></row><row><cell>Word2vec-sbwc</cell><cell>300</cell><cell>0.795</cell><cell>0.624</cell></row><row><cell>n2v-Inverse Association</cell><cell>300</cell><cell>0.837</cell><cell>0.650</cell></row><row><cell>n2v-Inverse Association</cell><cell>200</cell><cell>0.844</cell><cell>0.664</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_0">https://github.com/jocarema/nalc_vectors</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_1">https://github.com/jocarema/nap_vectors</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="6" xml:id="foot_2">http://www.jeuxdemots.org/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="7" xml:id="foot_3">https://smallworldofwords.org/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="8" xml:id="foot_4">http://www.eat.rl.ac.uk/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="9" xml:id="foot_5">http://web.usf.edu/FreeAssociation</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="10" xml:id="foot_6">http://snap.stanford.edu/node2vec/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="11" xml:id="foot_7">http://web.eecs.umich.edu/ ~mihalcea/downloads.html</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="12" xml:id="foot_8">https://github.com/uchile-nlp/spanish-word-embeddings</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="13" xml:id="foot_9">https://code.google.com/archive/p/word2vec/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="14" xml:id="foot_10">https://nlp.stanford.edu/projects/glove/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="15" xml:id="foot_11">https://github.com/facebookresearch/fastText/blob/master/ pretrained-vectors.md</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgments</head><p>This work was partially supported by the following projects: Conacyt FC-2016-01-2225 and PAPIIT IA401219, IN403016, AG400119.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<title level="m" type="main">Words in the mind: An introduction to the mental lexicon</title>
		<author>
			<persName><forename type="first">J</forename><surname>Aitchison</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2012">2012</date>
			<publisher>John Wiley &amp; Sons</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<monogr>
		<title level="m" type="main">Corpus de normas de asociación de palabras para el espaol de Mxico</title>
		<author>
			<persName><forename type="first">N</forename><surname>Arias-Trejo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">B</forename><surname>Barrón-Martínez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">H L</forename><surname>Alderete</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">A R</forename><surname>Aguirre</surname></persName>
		</author>
		<imprint>
			<publisher>NAP</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Don&apos;t count, predict! a systematic comparison of context-counting vs. context-predicting semantic vectors</title>
		<author>
			<persName><forename type="first">M</forename><surname>Baroni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Dinu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Kruszewski</surname></persName>
		</author>
		<ptr target="http://www.aclweb.org/anthology/P14-1023" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 52 nd Annual Meeting of the Association for Computational Linguistics</title>
		<title level="s">Long Papers</title>
		<meeting>the 52 nd Annual Meeting of the Association for Computational Linguistics</meeting>
		<imprint>
			<date type="published" when="2014">2014</date>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page" from="238" to="247" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Wan2vec: Embeddings learned on word association norms</title>
		<author>
			<persName><forename type="first">G</forename><surname>Bel-Enguix</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Gómez-Adorno</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Reyes-Magaña</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Sierra</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Semantic Web</title>
				<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<monogr>
		<title level="m" type="main">Enriching word vectors with subword information</title>
		<author>
			<persName><forename type="first">P</forename><surname>Bojanowski</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Grave</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Joulin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Mikolov</surname></persName>
		</author>
		<idno type="DOI">10.1162/tacla00051</idno>
		<idno type="arXiv">arXiv:1607.04606</idno>
		<ptr target="https://arxiv.org/abs/1607.04606" />
		<imprint>
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
	<note type="report_type">Computing Research Repository</note>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Navigating word association norms to extract semantic information</title>
		<author>
			<persName><forename type="first">J</forename><surname>Borge-Holthoefer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Arenas</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 31 st Annual Conference of the Cognitive Science Society</title>
				<meeting>the 31 st Annual Conference of the Cognitive Science Society</meeting>
		<imprint>
			<date type="published" when="2009">2009</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<monogr>
		<author>
			<persName><forename type="first">C</forename><surname>Cardellino</surname></persName>
		</author>
		<ptr target="http://crscardellino.github.io/SBWCE/" />
		<title level="m">Spanish Billion Words Corpus and Embeddings</title>
				<imprint>
			<date type="published" when="2016-03">March 2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Associative strength and semantic activation in the mental lexicon: Evidence from continued word associations</title>
		<author>
			<persName><forename type="first">S</forename><surname>De Deyne</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">J</forename><surname>Navarro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Storms</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 35 th Annual Conference of the Cognitive Science Society</title>
				<meeting>the 35 th Annual Conference of the Cognitive Science Society</meeting>
		<imprint>
			<publisher>Cognitive Science Society</publisher>
			<date type="published" when="2013">2013</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Predicting human similarity judgments with distributional models: The value of word associations</title>
		<author>
			<persName><forename type="first">S</forename><surname>De Deyne</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Perfors</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">J</forename><surname>Navarro</surname></persName>
		</author>
		<idno type="DOI">10.24963/ijcai.2017/671</idno>
		<ptr target="https://doi.org/10.24963/ijcai.2017/671" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of COLING 2016, the 26 th International Conference on Computational Linguistics: Technical Papers</title>
				<meeting>COLING 2016, the 26 th International Conference on Computational Linguistics: Technical Papers</meeting>
		<imprint>
			<date type="published" when="2016">2016</date>
			<biblScope unit="page" from="1861" to="1870" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<monogr>
		<author>
			<persName><forename type="first">A</forename><surname>Fernandez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Díez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Alonso</surname></persName>
		</author>
		<ptr target="http://inico.usal.es/usuarios/gimc/normas/index_nal.asp" />
		<title level="m">Normas de asociación libre en castellano de la universidad de salamanca</title>
				<imprint>
			<date type="published" when="2010">2010</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Placing search in context: The concept revisited</title>
		<author>
			<persName><forename type="first">L</forename><surname>Finkelstein</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Gabrilovich</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Matias</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Rivlin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Solan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Wolfman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Ruppin</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 10 th International Conference on World Wide Web</title>
				<meeting>the 10 th International Conference on World Wide Web</meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2001">2001</date>
			<biblScope unit="page" from="406" to="414" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Improving feature representation based on a neural network for author profiling in social media texts</title>
		<author>
			<persName><forename type="first">H</forename><surname>Gómez-Adorno</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Markov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Sidorov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Posadas-Durán</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">A</forename><surname>Sanchez-Perez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Chanona-Hernandez</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Computational Intelligence and Neuroscience</title>
		<imprint>
			<biblScope unit="page">13</biblScope>
			<date type="published" when="2016">2016. 2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Document embeddings learned on various types of n-grams for cross-topic authorship attribution</title>
		<author>
			<persName><forename type="first">H</forename><surname>Gómez-Adorno</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">P</forename><surname>Posadas-Durán</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Sidorov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Pinto</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Computing</title>
		<imprint>
			<biblScope unit="page" from="1" to="16" />
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">node2vec: Scalable feature learning for networks</title>
		<author>
			<persName><forename type="first">A</forename><surname>Grover</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Leskovec</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 22 nd ACM International Conference on Knowledge Discovery and Data Mining</title>
				<meeting>the 22 nd ACM International Conference on Knowledge Discovery and Data Mining</meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2016">2016</date>
			<biblScope unit="page" from="855" to="864" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Cross-lingual semantic relatedness using encyclopedic knowledge</title>
		<author>
			<persName><forename type="first">S</forename><surname>Hassan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Mihalcea</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2009 Conference on Empirical Methods in Natural Language Processing</title>
				<meeting>the 2009 Conference on Empirical Methods in Natural Language Processing</meeting>
		<imprint>
			<date type="published" when="2009">2009</date>
			<biblScope unit="volume">3</biblScope>
			<biblScope unit="page" from="1192" to="1201" />
		</imprint>
	</monogr>
	<note>Association for Computational Linguistics</note>
</biblStruct>

<biblStruct xml:id="b15">
	<monogr>
		<author>
			<persName><forename type="first">D</forename><surname>Jackson-Maldonado</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Thal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Fenson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Marchman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Newton</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Conboy</surname></persName>
		</author>
		<title level="m">Macarthur inventarios del desarrollo de habilidades comunicativas (inventarios): Users guide and technical manual</title>
				<meeting><address><addrLine>Baltimore, MD</addrLine></address></meeting>
		<imprint>
			<publisher>Brookes</publisher>
			<date type="published" when="2003">2003</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<monogr>
		<title level="m" type="main">An associative thesaurus of English and its computer analysis</title>
		<author>
			<persName><forename type="first">G</forename><surname>Kiss</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Armstrong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Milroy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Piper</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1973">1973</date>
			<publisher>Edinburgh University Press</publisher>
			<pubPlace>Edinburgh</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Making people play for lexical acquisition</title>
		<author>
			<persName><forename type="first">M</forename><surname>Lafourcade</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the th SNLP 2007, Pattaya</title>
				<meeting>the th SNLP 2007, Pattaya</meeting>
		<imprint>
			<date type="published" when="2007-12">December 2007</date>
			<biblScope unit="volume">7</biblScope>
			<biblScope unit="page" from="13" to="15" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<monogr>
		<title level="m" type="main">Efficient estimation of word representations in vector space</title>
		<author>
			<persName><forename type="first">T</forename><surname>Mikolov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Corrado</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Dean</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1301.3781</idno>
		<ptr target="https://arxiv.org/abs/1301.3781" />
		<imprint>
			<date type="published" when="2013">2013</date>
		</imprint>
	</monogr>
	<note type="report_type">Computing Research Repository</note>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">Contextual correlates of semantic similarity</title>
		<author>
			<persName><forename type="first">G</forename><surname>Miller</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Charlees</surname></persName>
		</author>
		<idno type="DOI">10.1080/01690969108406936</idno>
		<ptr target="https://doi.org/10.1080/01690969108406936" />
	</analytic>
	<monogr>
		<title level="j">Language and cognitive processes</title>
		<imprint>
			<biblScope unit="volume">6</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="1" to="28" />
			<date type="published" when="1991">1991</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<monogr>
		<title level="m" type="main">Word association rhyme and word fragment norms</title>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">L</forename><surname>Nelson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">L</forename><surname>Mcevoy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">A</forename><surname>Schreiber</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1998">1998</date>
			<publisher>The University of South Florida</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">The distributional hypothesis</title>
		<author>
			<persName><forename type="first">M</forename><surname>Sahlgren</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Italian Journal of Disability Studies</title>
		<imprint>
			<biblScope unit="volume">20</biblScope>
			<biblScope unit="page" from="33" to="53" />
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">Evaluation methods for unsupervised word embeddings</title>
		<author>
			<persName><forename type="first">T</forename><surname>Schnabel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Labutov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Mimno</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Joachims</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing</title>
				<meeting>the 2015 Conference on Empirical Methods in Natural Language Processing</meeting>
		<imprint>
			<date type="published" when="2015">2015</date>
			<biblScope unit="page" from="298" to="307" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<monogr>
		<title level="m" type="main">Word association thesaurus as a resource for extending semantic networks</title>
		<author>
			<persName><forename type="first">A</forename><surname>Sinopalnikova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Smrz</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2004">2004</date>
			<biblScope unit="page" from="267" to="273" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<analytic>
		<title level="a" type="main">Conceptual graphs as a universal knowledge representation</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">F</forename><surname>Sowa</surname></persName>
		</author>
		<idno type="DOI">10.1016/0898-1221</idno>
		<idno>(92)90137-7</idno>
		<ptr target="https://doi.org/10.1016/0898-1221" />
	</analytic>
	<monogr>
		<title level="j">Computers &amp; Mathematics with Applications</title>
		<imprint>
			<biblScope unit="volume">23</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page" from="75" to="93" />
			<date type="published" when="1992">1992</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
