<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Word Similarity Perception: an Explorative Analysis</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Alice</forename><surname>Ruggeri</surname></persName>
							<email>ruggeri@di.unito.it</email>
							<affiliation key="aff0">
								<orgName type="department">Centre for Cognitive Science</orgName>
								<orgName type="institution">University of Turin</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Loredana</forename><surname>Cupi</surname></persName>
							<email>loredana.cupi@unito.it</email>
							<affiliation key="aff1">
								<orgName type="department">Department of Foreign Languages and Literatures and Modern Cultures</orgName>
								<orgName type="institution">University of Turin</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Luigi</forename><forename type="middle">Di</forename><surname>Caro</surname></persName>
							<email>dicaro@di.unito.it</email>
							<affiliation key="aff2">
								<orgName type="department">Department of Computer Science</orgName>
								<orgName type="institution">University of Turin</orgName>
							</affiliation>
						</author>
						<title level="a" type="main">Word Similarity Perception: an Explorative Analysis</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">B449B736B982CDD7E5DF40150EC3BC3A</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T19:19+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Natural language is a medium for expressing things belonging to conceptual and cognitive levels, made of words and grammar rules used to carry semantics. However, its natural ambiguity is the main critical issue that computational systems are generally asked to solve. In this paper, we propose to go beyond the current conceptualization of word similarity, i.e., the building block of disambiguation at computational level. First, we analyze the origin of the perceived similarity, studying how conceptual, functional, and syntactic aspects influence its strength. We report the results of a two-stages experiment showing clear similarity perception patterns. Then, based on the insights gained in the cognitive tests, we developed a computational system that automatically predicts word similarity reaching high levels of accuracy.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Introduction</head><p>Words are symbolic entities referring to something which fills a portion of an autopoietic space made of conceptual, cognitive and contextual information. These three aspects are fundamental to understand the meaning ascribed to linguistic expressions.</p><p>One of the most important building block in almost all Computational Linguistics tasks is the computation of similarity scores between texts at different levels: words, sentences and discourses. Since manual numeric annotations of word-word similarity revealed a low agreement between the annotators, cognitive studies can help improve computational systems by discovering what lies behind the perception of similairty between words and their referenced concepts. The concept of similarity has been extensively studied in the Cognitive Science community, since it is fundamental in the human cognition. We tend to rely on similarity to generate inferences and categorize objects into kinds when we do not know exactly what properties are relevant, or when we cannot easily separate an object into separate properties. when specific knowledge is available, then a generic assessment of similarity is less relevant (G. L. <ref type="bibr">Murphy &amp; Medin,1985)</ref>.</p><p>Since words co-occur in textual representations (mutually influencing one each other) it is possible to make experiments on contextual information to analyze what and how influence the perception of similarity. Let us consider two words as two mental representations. The intersection between them can be seen as the context that may also help define the correct similarity of the words in a text. For example, sugar and salt can be easily associated to the context kitchen, whereas salt and sea intersect in another part of the mental representations.</p><p>From a computational perspective, being words ambiguous by nature, the disambiguation process (i.e., Word Sense Disambiguation) is one of the most studied tasks in Computational Linguistics. To make an example, the term count can mean many things like nobleman or sum. Using contextual information, it is often possible to make a choice. Again, this choice is done by means of comparisons among contexts, that are still made of words. In other terms, we may state that the computational part of almost all computational linguistics research is about the calculus of matching scores between linguistic items, i.e., words similarity. But what's behind words similarity?</p><p>There exist many annotated data related to similarity and relatedness between words, like wordsim-353 <ref type="bibr" target="#b5">(Finkelstein et al.,2001)</ref> and SimLex-999 <ref type="bibr">(Hill, Reichart, &amp; Korhonen,2014)</ref>. A large part of the proposed computational systems aims at finding relatedness between words, instead of similarity. Relatedness is more general than similarity, since it refers to a generic correlation (like cradle and baby, that are words representing dissimilar concepts which, however, share similar contexts).</p><p>One problem of similarity, as often faced in literature or annotated in datasets, is that it cannot be a static value. Indeed, as the authors of these resources state in their works, the agreement between the annotators is usually not high (around 50-70%). The reason is trivial, however: people can give different degrees of importance with respect to the specific characteristics of the concepts to compare. If we ask one to say how much dog is similar to cat, the right answer can only be "it depends". While we can all agree about the fact that the concept dog is quite similar to cat, we cannot say 0.7 rather than 0.9 (in the range [0,1]) with certainty. Different aspects can be taken into account: are we measuring the form of the animal, or its behaviour? In both cases, it depends on which part of the animal and which actions we are considering to make a choice. For instance, dogs use to return thrown objects. From this point of view, dogs and cats are dissimilar.</p><p>In the light of this, our contribution provides the basis for understanding what lies behind a similarity between words and their referenced concepts. First, we analyze syntactic, conceptual and functional aspects of the similarity perception; then, we develop a computational system which is able to predict similarity by leveraging contextual information.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>The Cognitive Experiment</head><p>In this work, we present two tests to analyze how linguistic constructions are perceived by humans in terms of strength of semantic similarity and if there exists a functionality-based connection that has an influence on its perception. The experiment was presented to 96 users, having different ages and professions, without any particular cognitive or linguistic disorder.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Test on single words</head><p>The first test of the experiment regards the perception of the similarity between single words<ref type="foot" target="#foot_0">1</ref> . In particular, the goal was to analyze how the users focus on the functional links between the words, and more importantly if such functionalbased similarity is a preferential perception channel compared to the conceptual-based one.</p><p>Words are ambigous, and many resources have been released with the goal of defining all the possible senses of a word (i.e., WordNet). Word Sense Disambiguation <ref type="bibr">(Bhattacharyya &amp; Khapra,2012)</ref> is the task of resolving the ambiguity of a word in a given context. Notice that, in our experiment, we do not need any disambiguation of the words, since this process is embodied in the human cognition, thus the users of the test will autonomously represent their subjective sense to associate to the words under comparison.</p><p>Then, since we wanted to compare conceptual with functional preferences, we designed the test as a comparison between two word pairs, one involving conceptually-related words and one with words linked by direct functionalities. To generalize, let us consider the words a, b, and c with the conceptual word pair a-b and the functional word pair a-c. The user is asked to mark the most similar word (among b and c) to associate to a, and so the most correlated word pair. The users were not aware of the goal of the test and of the difference between the word pairs. Since words and actions present a high variability in terms of conceptual range (or their mental representation), we put particular attention to the choice of the word pairs, according to the following principles:</p><p>Conceptual granularity If we think at the words object and thing, we probably do not have enough information to make significant comparisons due to their large and undefined conceptual boundaries. The same happens in cases when two words represent very specific concepts such as lactose and amino acid. The word pairs of the proposed test have been selected by considering this constraint (and so they include words which are not too specific nor too general).</p><p>Concreteness Words may have direct links with concrete objects such as "table" and "dog". In other cases, words such as "justice" and "thought" represent abstract concepts. Since it is not clear how this may affect the per-ception of similarity, we decided to keep concrete words only.</p><p>Semantic coherence Another criterion used for the selection of the words was the level of semantic similarity between the word pairs to compare. To better analyze whether the functional aspect plays a significant role in the similarity perception, we extracted conceptual and functional pairs of words which had similar semantic closeness according to a standard semantic similarity calculation. In the light of this, we used a Latent Semantic Space calculated over almost 1 million of documents coming from the collection of literary text contained in the project Gutenberg page<ref type="foot" target="#foot_1">2</ref> . The selected conceptual and functional word pairs had the property of having a very close semantic similarity (the score differences were less than 0,01 in a [0,1] range).</p><p>The test was composed by three word pairs, to leave it simple and to be not affected by users tiredness. Then, instead of randomly selecting three different word pairs, we wanted to consider three cases in which the functional links between the words have distinct levels of importance. Our assumption was that the more the importance of the functional link between two words in a pair, the more its perceived similarity (and thus the user preferences with respect to the conceptual word pair). For this reason we added a final criterion:</p><p>Increasing relevance of the functional aspect To estimate the importance of the functional aspect that relates two words we analyzed the number of actions (or verbs) in which they are usually involved with. In our test, the functional word pairs salt-water, nail-polish, and ring-finger have a functional link of 0.003<ref type="foot" target="#foot_2">3</ref> , 0.0101<ref type="foot" target="#foot_3">4</ref> , and 0.0625<ref type="foot" target="#foot_4">5</ref> respectively (see Table <ref type="table" target="#tab_0">1</ref>). These values are calculated in the following way: given the total number of existing verbs NV(rw) for the root word rw and the number of effective usages EU(sw) with the second word of the pair sw, we computed the functional link Fl(rw, sw) of the functional pair as EU(sw) / NV(rw). We then considered a backup test set with random word pairs matching with the same above-mentioned criteria, collecting a total of 24 answers on 24 different cases of word pairs such as the main one of Table <ref type="table" target="#tab_0">1</ref>. This was done to prove the reliability of the test, seeing whether the results and the analyses show a similar trend, being independent from the selection of the words. The results of the whole test is described in the final part of this section.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Test on Phrases</head><p>The second test of the experiment concerns the perception of the similarity between phrases, o multi-word linguistic constructions <ref type="foot" target="#foot_5">6</ref> . The goal was to analyze how the syntactic context of a target word influences the perception of similarity among entire phrases. More in detail, we wanted to discover possible differences of such perception along different syntactic roles. We considered a simple syntactic structure of the type subject-verb-object.</p><p>Given a root sentence such as "Mario sings the song", we created three variations by changing the subject, the verb, and the direct object. For example, by changing "Mario" with "The bird" we obtained "The bird sings the song". The complete set of replacements are shown in Table <ref type="table" target="#tab_2">3</ref>.</p><p>We presented to the 96 users a total of 4 sentences (see Table 2), that with its 4 variations produce a total of 16 pairs of sentences to be analyzed by the users in terms of perceived similarity, as in the first test. For each sentence, the users had to indicate the degree of similarity of the original sentence with one of its variation using a value in the range [0,10]<ref type="foot" target="#foot_6">7</ref> . 0 means no semantic similarity between the two phrases and 10 means total equality. The grammatical changes made on the original sentences were chosen maintaining the semantic validity (i.e., all the sentences represent valid mental representations). </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Interpretation of the results</head><p>In this section, we give a preliminary interpretation of the results on the collected answers.</p><p>In the first test on single words, we can state that, generally, conceptual and functional word pairs are differently perceived according to the importance of the funcional link in the functional word pair. This shows that words and their referenced concepts are mainly compared in terms of conceptual similarity, but when there exists importafnt functionalities between them, this influences the users preference towards the functional word pair ??.</p><p>For example, the similarity of the sugar-salt pair results to be stronger compared to the water-salt one, since the action to add/put the salt in the water is "a needle in a haystack" with respect to all the actions related to water and salt independently. This means that there is no exclusive action between water and salt (i.e., there are many actions that involve water). An opposite example is represented by the word pair ring-finger, since the action to put/wear the ring on the finger is much more exclusive than in the previous case. Such preference could be explained by stating that all word pairs, especially with words that underlie actions, have a strong visual representation that makes them quickly perceivable. This result is also in line with what stated by <ref type="bibr" target="#b4">(Cohen et al.,2002)</ref>, i.e., words that have a functionality-based relationship can have a more complex visual component that makes such correlation weaker.</p><p>In Figure <ref type="figure" target="#fig_0">1</ref>, we show the users preferences for the second test. In the case of verb replacement (VC) we can notice a high meaning change in terms of similarity perception (similarity values close to 0), so the verb represents the real root of the mental representations. The case of the subject change (SC) shows a less important decrease of similarity perception, while the object change (OC) resulted to be the less relevant syntactic role influencing the meaning of the whole phrase.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>The Computational Analysis</head><p>In the previous section we studied the role of the context (on different levels) within the process of word similarity perception. Since the results indicated that both functional aspects and syntactic roles have an impact on how people perceive similarity, we experimented a computational approach for the automatic estimation of the similarity based on functional and syntax-aware contextual information.</p><p>In particular, we used the large and freely-available semantic resource ConceptNet<ref type="foot" target="#foot_7">8</ref> . A partial overview of the semantic knowledge contained in ConceptNet is illustrated in Table <ref type="table" target="#tab_5">6</ref>. ConceptNet is a resource based on common-sense rather than linguistic knowledge since it contains much more functionbased information (e.g., all the actions an object can or cannot do) contained in even complex syntactic structures. The idea is also to exploit users perception of reality (the actual origin of ConceptNet) instead of the result of top-down expert building of ontologies (e.g., WordNet). ConceptNet contains important semantic problems related to covarage, utility of semantic information and coherence, but we used it as a black box due to its largeness and common-sense nature. A deep analysis of this resource is out of the scope of this paper.</p><p>The experiment started from the transformation of a wordword-score similarity dataset into a context-based dataset in which the words are replaced by sets of semantic information taken from ConceptNet. The aim was to figure out which semantic facts make the similarity between two words perceivable.</p><p>We used the dataset SimLex-999 <ref type="bibr" target="#b7">(Hill et al.,2014)</ref> that contains one thousand word pairs that were manually annotated with similarity scores. The inter-annotation agreement is 0.67 (Spearman correlation). We leveraged ConceptNet to retrieve the semantic information associated to the words of each pair, then keeping the intersection. For example, considering the pair rice-bean, ConceptNet returns the following set of semantic information for the term rice:</p><p>[hasproperty-edible, isa-starch, memberof-oryza, atlocation-refrigerator, usedfor-survival, atlocationatgrocerystore, isa-food, isa-domesticateplant, relatedto-grain, madeof-sake, isa-grain, receivesactioncook, atlocation-pantry, atlocation-ricecrisp, atlocationsupermarket, ...] Then, the semantic information for the word bean are:</p><p>[usedfor-fillbeanbagchair, atlocation-infield, atlocation-can, usedfor-nutrition, usedfor-cook, atlocation-atgrocerystore, usedfor-grow, atlocationfoodstore, isa-legume, usedfor-count, isadomesticateplant, atlocation-cookpot, atlocationbeansoup, atlocation-soup, isa-vegetable, ...] Finally, the intersection produces the following set:</p><p>[atlocation-atgrocerystore, isa-domesticateplant, atlocationpantry] At this point, for each non-empty intersection, we created one instance of the type: &lt;semantic information&gt;, &lt;similarity score&gt; and computed a standard term-document matrix, where the term is a semantic term within the set of semantic information retrieved from ConceptNet and the document dimension represents the word pairs of the original dataset. After this preprocessing phase, the score attribute is discretized into two bins:</p><p>• non-similar class -range in the dataset [0, 5]</p><p>• similar class -range in the dataset <ref type="bibr">[5.1, 10]</ref> The splitting of the data into two clusters allowed us to experiment a classic supervised classification system, where a Machine Learning tool (a Support Vector Machine, in our case) has been used to learn a binary model for automatically classifying similar and non-similar word pairs. The result of the experiment is shown in Table <ref type="table" target="#tab_6">7</ref>. Noticeably, the classifier has been able to reach a quite good accuracy (65.38% of correctly classified word pairs), considering that the interannotation agreement of the original data is only 0.67 (Spearman correlation). Notice that similar word pairs are generally easier to identify with respect to non-similar ones.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Related Work</head><p>This paper presents an idea which combines linguistic, cognitive and computational perspectives. In this section, we mention those theoretical and empirical methods that inspired our motivational basis.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Linguistic Background</head><p>The difficulty of defining the meaning of meaning has to do with some tricky issues like lexical ambiguity and polysemy, vagueness, contextual variability of word meaning, etc. As a matter of fact, words are organized in lexicon as a complex network of semantic relation which are basically subsumed within Saussure's paradigmatic (the axis of combination) and syntagmatic (the axis of choice) axes <ref type="bibr">(Saussure,1983)</ref>. Some authors <ref type="bibr">(Chaffin &amp; Herrmann,1984)</ref> have already suggested theoretical and empirical taxonomies of semantic relations consisting of some main families of relation (such as contrast, similars, class inclusion, part-whole, etc.). As Murphy points out (M. L. <ref type="bibr">Murphy,2003)</ref>, lexicon has become more central in linguistic theories and, even if there is no a widely accepted theory on its internal semantic structure and how lexical information are represented in it, the semantic relations among words are considered in scholarly literature as relevant to the structure of both lexical and conceptual information and it is generally believed that relations among words determine meaning.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Cognitive Background</head><p>Although words perception could seem immediate, the input we perceive is recognized and trasformed mediating background and contextual information, within a dynamic and cooperative process. The well-known semiotic triangle <ref type="bibr">(Ogden, Richards, Malinowski, &amp; Crookshank,1946)</ref> introduced by different authors over time represents a first reference for our study. People use symbols (our words) to communicate meanings (the effective content). The meaning is something untangible, which can be though even without any concrete presence. The last point is then the physical reference, i.e., the object in the reality<ref type="foot" target="#foot_8">9</ref> . Note that there is no connection between symbols and references, since only imagined meanings can allow the two to be linked.</p><p>Interaction is another important aspect that has been investigated in literature. Indeed, the actions change the type of perception of an object, which models itself to fit with the context of use. Then, the Gestalt theory <ref type="bibr">(Köhler,1929)</ref> contains different notions about the perception of meaning according to interaction and context. In particular, the core of the model is the complementarity between the figure and the ground. In our case, a word is the figure and the ground is the context that lets emerge its specific sense. Finally, James Gibson introduced the concept of affordances as the cognitive cues that an object exposes to the external world, indicating ways of use <ref type="bibr">(Gibson,1977)</ref>. In cognitive and computational linguistics, this theory can be inherited to model words as objects and contexts as their interaction with the world.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Computational Background</head><p>In this section, we review the main works that are related to our contribution from a computational perspective. Natural Language Processing represents an active research community whose focus is letting machines communicate by understanding semantics within linguistic expressions. Ontology Learning <ref type="bibr">(Cimiano,2006)</ref> is the task of automatic extracting structured semantic knowledge from texts, and it well fits the scope of this paper. Nevertheless, Word Sense Disambiguation (WSD) <ref type="bibr">(Stevenson &amp; Wilks,2003)</ref> is maybe the most related NLP task, whose aim is to capture the correct meaning of a word in a context. Generally speaking, many other tasks have the problem of comparing linguistic items in order to make choices to pass from syntax to semantics. Named Entity Recognition (NER) <ref type="bibr">(Nadeau &amp; Sekine,2007;</ref><ref type="bibr">Marrero, Urbano, Sánchez-Cuadrado, Morato, &amp; Gómez-Berbís,2013)</ref> is the task of identifying entities like people, organizations and locations in texts. This is often done by comparing words in contexts to some learned patterns. In general, many other NLP tasks are based on the evaluation of similarity scores <ref type="bibr">(Manning &amp; Schütze,1999)</ref>.</p><p>Nowadays, there exists a large set of available semantic resources that can be used in Natural Language Processing techniques in order to understand the hidden meaning of perceived similarity between two words or concepts. For example, ConceptNet contains semantic information that are usually associated with common terms (even if not correctly disambiguated). By analyzing the relationship betweeen annotated similarity scores and semantic information it is possi-ble to create predictive models which automatically deduce words similarity by dynamically weighting words features based on their mutual interaction.</p><p>If we consider the objects / agents / actions to be terms in text sentences, we can try to extract their meaning and semantic constraints by using the idea of affordances. For instance, let us think to the sentence "The squirrel climbs the tree". In this case, we need to know what kind of subject "squirrel" is to figure out (and visually imagine) how the action will be performed. According to this, no particular issues come out from the reading of this sentence. Let us now consider the sentence "The elephant climbs the tree". Even if the grammatical structure of the sentence is the same as before, the agent of the action is different, and it obviously creates some semantic problems. In fact, from this case, some constraints arise; in order to climb a tree, the subject needs to fit to our mental model of something that can climb a tree. In addition, this also depends on the mental model of "tree". Moreover, different agents can be both correct subjects of an action whilst they may produce different meanings in terms of how the action will be mentally performed. Consider the sentences "The cat opens the door" and "The man opens the door". In both cases, some implicit knowledge suggests the manner the action is done: while in the second case we may think at the cat that opens the door leaning to it, in the case of the man we probably imagine the use of a door handle. A study of these language dynamics can be of help for many NLP tasks like Part-Of-Speech tagging as well as more complex operations like dependency parsing and semantic relations extraction. Some of these concepts are latently studied in different disciplines related to statistics. Distributional Semantics (DS) <ref type="bibr">(Baroni &amp; Lenci,2010)</ref> represents a class of statistical and linguistic analysis of text corpora that tries to estimate the validity of connections between subjects, verbs, and objects by means of statistical sources of significance.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Conclusions and Future Work</head><p>In this paper, we proposed a combined analysis of linguistic, cognitive and computational aspects to assess the nature of words similarity. First, we studied how word similarity perception is influenced in terms of conceptual, functional and syntactic roles. In future work, our aim is to extend the sample of users on more specific cases. Still, by changing the language of the users, we can have results that take into account the cultural ground, understanding if and how word similarity depends on it. On the other side, we stressed the importance of computational understanding of similarity to improve Computational Linguistics tasks which are based on it, usually without any analysis of contextual information. In particular, we used the large semantic knowledge contained in ConceptNet to create a Support Vector Machine classifier to predict word similarity based on an annotated dataset. In future work, we will extend our experimental analysis to validate existing similarity datasets and to produce predictive models for the automatic identification of human-readable similarity scores.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Results of the second test, showing the change scores in terms of word similairty perception after subject, verb and object replacements. SC stands subject change, VC for verb change, and OC for object change.</figDesc><graphic coords="4,62.15,75.64,226.48,136.37" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1 :</head><label>1</label><figDesc>The chosen word pairs in the first test.</figDesc><table><row><cell cols="3">Root word Conceptual Pair Functional Pair [F. link]</cell></row><row><cell>rw</cell><cell>rw -sw</cell><cell>rw -sw [Fl(rw,sw)]</cell></row><row><cell>salt</cell><cell>salt -sugar</cell><cell>salt -water [0.003]</cell></row><row><cell>nail</cell><cell>nail -finger</cell><cell>nail -polish [0.0101]</cell></row><row><cell>ring</cell><cell>ring -necklace</cell><cell>ring -finger [0.0625]</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2 :</head><label>2</label><figDesc>The chosen phrases in the second test.</figDesc><table><row><cell>Phrase ID</cell><cell>Phrase</cell></row><row><cell>(a)</cell><cell>Mario sings the song</cell></row><row><cell>(b)</cell><cell>Alan drives the car</cell></row><row><cell>(c)</cell><cell>Alice writes the book</cell></row><row><cell>(d)</cell><cell>Marco does the homeworks</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 3 :</head><label>3</label><figDesc>The word replacements for subjects (SC), verbs (VC) and direct objects (OC).</figDesc><table><row><cell>Replacement</cell><cell>(a)</cell><cell>(b)</cell><cell>(c)</cell><cell>(d)</cell></row><row><cell>SC</cell><cell>bird</cell><cell cols="3">robot computer software</cell></row><row><cell>VC</cell><cell cols="2">writes cleans</cell><cell>cleans</cell><cell>gives</cell></row><row><cell>OC</cell><cell>verse</cell><cell>band</cell><cell>sheet</cell><cell>pasta</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 4 :</head><label>4</label><figDesc>Results showing the percentage of preferences in the choice of the most (perceived) correlated word pairs of the first test.</figDesc><table><row><cell>Case</cell><cell>Word pairs</cell><cell>N. of preferences</cell><cell>%</cell></row><row><cell>1a.</cell><cell>salt -sugar</cell><cell>75</cell><cell>78%</cell></row><row><cell>1b.</cell><cell>salt -water</cell><cell>21</cell><cell>22%</cell></row><row><cell>2a.</cell><cell>nail -finger</cell><cell>44</cell><cell>46%</cell></row><row><cell>2b.</cell><cell>nail -polish</cell><cell>52</cell><cell>54%</cell></row><row><cell>3a.</cell><cell>ring -necklace</cell><cell>16</cell><cell>17%</cell></row><row><cell>3b.</cell><cell>ring -finger</cell><cell>80</cell><cell>83%</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_4"><head>Table 5 :</head><label>5</label><figDesc>Results on the backup of the first test (with 24 different cases including 8 low-FL cases, 8 medium-FL cases and 8 high-FL cases). The results are in line with the ones of the main test shown in Table4.</figDesc><table><row><cell>Funct. Pair</cell><cell>Pref. w.r.t conceptual pair</cell></row><row><cell>Funct. Pair (low FL)</cell><cell>1 out of 8 (12.5%)</cell></row><row><cell>Funct. Pair (medium FL)</cell><cell>5 out of 8 (62.5%)</cell></row><row><cell>Funct. Pair (high FL)</cell><cell>7 out of 8 (87.5%)</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_5"><head>Table 6 :</head><label>6</label><figDesc>Some of the existing relations in ConceptNet, with example sentences in English.</figDesc><table><row><cell>Relation</cell><cell>Example sentence</cell></row><row><cell>IsA</cell><cell>NP is a kind of NP.</cell></row><row><cell>LocatedNear</cell><cell>You are likely to find NP near NP.</cell></row><row><cell>UsedFor</cell><cell>NP is used for VP.</cell></row><row><cell>DefinedAs</cell><cell>NP is defined as NP.</cell></row><row><cell>HasA</cell><cell>NP has NP.</cell></row><row><cell>HasProperty</cell><cell>NP is AP.</cell></row><row><cell>CapableOf</cell><cell>NP can VP.</cell></row><row><cell>ReceivesAction</cell><cell>NP can be VP.</cell></row><row><cell>HasPrerequisite</cell><cell>NP-VP requires NP-VP.</cell></row><row><cell cols="2">MotivatedByGoal You would VP because you want VP.</cell></row><row><cell>MadeOf</cell><cell>NP is made of NP.</cell></row><row><cell>...</cell><cell>...</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_6"><head>Table 7 :</head><label>7</label><figDesc>Classification results in terms of Precision, Recall, and F-measure. The total accuracy is 65.38%.</figDesc><table><row><cell cols="3">Precision Recall F-Measure</cell><cell>Class</cell></row><row><cell>0,697</cell><cell>0,475</cell><cell>0,565</cell><cell>non-similar</cell></row><row><cell>0,633</cell><cell>0,815</cell><cell>0,713</cell><cell>similar</cell></row><row><cell>0,664</cell><cell>0,654</cell><cell>0,643</cell><cell>weighted total</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">Notice that "similarity between words" is intended as the similarity between the concepts they bring to mind.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_1">http://promo.net/pg/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_2">For the verbs "to put", "to add" and "to get"</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_3">For the verbs "to apply" and "to use"</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_4">For the verbs "to put" and "to wear"</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="6" xml:id="foot_5">Even in this case, "similarity" is intended as the similarity between the concepts related to the phrases.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="7" xml:id="foot_6">  7  We used a [0,10] range instead of a [0,1] range as in the previous test because it represents a more human-understandable and intuitive votation.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="8" xml:id="foot_7">http://conceptnet5.media.mit.edu/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="9" xml:id="foot_8">The existing terminology is quite varying: symbolthought/reference/referent (Aristotele); object-representationinterpretant (Peirce); signified-sign-referent (De Saussure)</note>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Distributional memory: A general framework for corpus-based semantics</title>
		<author>
			<persName><forename type="first">M</forename><surname>Baroni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Lenci</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Computational Linguistics</title>
		<imprint>
			<biblScope unit="volume">36</biblScope>
			<biblScope unit="issue">4</biblScope>
			<biblScope unit="page" from="673" to="721" />
			<date type="published" when="2010">2010</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Word sense disambiguation</title>
		<author>
			<persName><forename type="first">P</forename><surname>Bhattacharyya</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Khapra</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Emerging Applications of Natural Language Processing: Concepts and New Research</title>
				<imprint>
			<date type="published" when="2012">2012</date>
			<biblScope unit="volume">22</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">The similarity and diversity of semantic relations</title>
		<author>
			<persName><forename type="first">R</forename><surname>Chaffin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">J</forename><surname>Herrmann</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Memory &amp; Cognition</title>
		<imprint>
			<biblScope unit="volume">12</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page" from="134" to="141" />
			<date type="published" when="1984">1984</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<monogr>
		<title level="m" type="main">Ontology learning from text</title>
		<author>
			<persName><forename type="first">P</forename><surname>Cimiano</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2006">2006</date>
			<publisher>Springer</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Language-specific tuning of visual cortex? functional properties of the visual word form area</title>
		<author>
			<persName><forename type="first">L</forename><surname>Cohen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Lehéricy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Chochon</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Lemer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Rivaud</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Dehaene</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Brain</title>
		<imprint>
			<biblScope unit="volume">125</biblScope>
			<biblScope unit="issue">5</biblScope>
			<biblScope unit="page" from="1054" to="1069" />
			<date type="published" when="2002">2002</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Placing search in context: The concept revisited</title>
		<author>
			<persName><forename type="first">L</forename><surname>Finkelstein</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Gabrilovich</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Matias</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Rivlin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Solan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Wolfman</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 10th international conference on world wide web</title>
				<meeting>the 10th international conference on world wide web</meeting>
		<imprint>
			<date type="published" when="2001">2001</date>
			<biblScope unit="page" from="406" to="414" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">The Theory of Affordances</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">J</forename><surname>Gibson</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Perceiving, acting, and knowing: Toward an ecological psychology</title>
				<editor>
			<persName><forename type="first">R</forename><surname>Shaw</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Bransford</surname></persName>
		</editor>
		<imprint>
			<publisher>Lawrence Erlbaum</publisher>
			<date type="published" when="1977">1977</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<monogr>
		<title level="m" type="main">Simlex-999: Evaluating semantic models with (genuine) similarity estimation</title>
		<author>
			<persName><forename type="first">F</forename><surname>Hill</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Reichart</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Korhonen</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1408.3456</idno>
		<imprint>
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b8">
	<monogr>
		<title level="m" type="main">Gestalt psychology</title>
		<author>
			<persName><forename type="first">W</forename><surname>Köhler</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1929">1929</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<monogr>
		<title level="m" type="main">Foundations of statistical natural language processing</title>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">D</forename><surname>Manning</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Schütze</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1999">1999</date>
			<publisher>MIT press</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Named entity recognition: Fallacies, challenges and opportunities</title>
		<author>
			<persName><forename type="first">M</forename><surname>Marrero</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Urbano</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Sánchez-Cuadrado</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Morato</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">M</forename><surname>Gómez-Berbís</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Computer Standards &amp; Interfaces</title>
		<imprint>
			<biblScope unit="volume">35</biblScope>
			<biblScope unit="issue">5</biblScope>
			<biblScope unit="page" from="482" to="489" />
			<date type="published" when="2013">2013</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">The role of theories in conceptual coherence</title>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">L</forename><surname>Murphy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">L</forename><surname>Medin</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Semantic relations and the lexicon</title>
				<editor>
			<persName><forename type="first">M</forename><forename type="middle">L</forename><surname>Murphy</surname></persName>
		</editor>
		<imprint>
			<publisher>Cambridge University Press</publisher>
			<date type="published" when="1985">1985. 2003</date>
			<biblScope unit="volume">92</biblScope>
			<biblScope unit="page">289</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">A survey of named entity recognition and classification</title>
		<author>
			<persName><forename type="first">D</forename><surname>Nadeau</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Sekine</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Lingvisticae Investigationes</title>
		<imprint>
			<biblScope unit="volume">30</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="3" to="26" />
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">The meaning of meaning</title>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">K</forename><surname>Ogden</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><forename type="middle">A</forename><surname>Richards</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Malinowski</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">G</forename><surname>Crookshank</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Harcourt</title>
				<meeting><address><addrLine>New York</addrLine></address></meeting>
		<imprint>
			<publisher>Brace &amp; World</publisher>
			<date type="published" when="1946">1946</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Embodied perception and the economy of action</title>
		<author>
			<persName><forename type="first">D</forename><surname>Proffitt</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Perspectives on psychological science</title>
		<imprint>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page" from="110" to="122" />
			<date type="published" when="2006">2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<monogr>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">D</forename><surname>Saussure</surname></persName>
		</author>
		<title level="m">Course in general linguistics, trans. R. Harris</title>
				<meeting><address><addrLine>London</addrLine></address></meeting>
		<imprint>
			<publisher>Duckworth</publisher>
			<date type="published" when="1983">1983</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Word-sense disambiguation</title>
		<author>
			<persName><forename type="first">M</forename><surname>Stevenson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Wilks</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">The Oxford Handbook of Comp. Linguistics</title>
				<imprint>
			<date type="published" when="2003">2003</date>
			<biblScope unit="page" from="249" to="265" />
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
