<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Measuring of Similarity between Pair of Words Using Word Net</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Atul</forename><surname>Gupta</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Department of Computer Science and Engineering</orgName>
								<orgName type="institution">Bhagwant University</orgName>
								<address>
									<settlement>Ajmer</settlement>
									<country key="IN">India</country>
								</address>
							</affiliation>
						</author>
						<author role="corresp">
							<persName><forename type="first">Kalpana</forename><surname>Sharma</surname></persName>
							<email>kalpanasharma56@gmail.com</email>
							<affiliation key="aff0">
								<orgName type="department">Department of Computer Science and Engineering</orgName>
								<orgName type="institution">Bhagwant University</orgName>
								<address>
									<settlement>Ajmer</settlement>
									<country key="IN">India</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Krishan</forename><forename type="middle">Kumar</forename><surname>Goel</surname></persName>
							<affiliation key="aff1">
								<orgName type="department">Department of Computer Application</orgName>
								<orgName type="institution">Raja Balwant Singh Management Technical Campus</orgName>
								<address>
									<settlement>Agra</settlement>
									<country key="IN">India</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Measuring of Similarity between Pair of Words Using Word Net</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">68C205A500FD366CD818FB9E29932843</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T00:47+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>WordNet</term>
					<term>Similarity Measure</term>
					<term>Semantic Similarity</term>
					<term>IS-A relationship</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>In the current era, the digital data size increased enormously and available abundantly.</p><p>Retrieving the relevant and accurate information from the available data still a big challenge. In this paper we are finding the similarity of noun-noun pairs and verb-verb pairs using Word-Net as corpus. Computation of similarity between noun-noun pairs in a sentence using different semantic algorithm computed and analysed.it has been observed that computation of similarity between verb-pair is found to be not as easy as computation of similarity between noun-pair. There are two challenges observed during the experimentation of this work. The first one is no standard data set available for verb pair and second is no exact hierarchy of verb of available in word-net.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Internet Searching has been integral part of life. There are lot of search engines available, but it has still some unsolved challenges like not able to provide accurate and exact search result. For example, if someone searches "Lincoln a Car " then it will be providing Car brand as well as it will provide output about Abraham Lincoln also. This search is not about the president of America President Abraham Lincoln, but the search engine will display these results also.</p><p>In the given example it is the Car and the Name Lincoln have been identified as noun-noun pair. So semantically it should relate the two nouns together and result will be retrieved as per the context given. Computation of similarity between word pair (noun-noun/ verb-verb) and sentence pair is still a huge problem for the researcher who work in the field of search engine, gene prioritization and NLP. Measuring of similarity between words is possible only in fixed domain for example: medical domain, engineering domain etc. Computing the similarity between noun pairs and verb pairs is done by lexical database. Words are connected in the form of lexical chain in lexical database i.e., Lucknow→City→Capital→Uttar Pradesh. Different semantic measure algorithms have been developed to compute the semantic closeness in the pair of words using the lexically connected database i.e., Word-Net. Words are present in hierarchal form in Word-Net. Various approaches have been implemented previously which uses lexical database as Word-Net. George A. <ref type="bibr">Miller [1]</ref> a psychology professor of Princeton university developed Word-Net in 1985.</p><p>Words are arranged in Word-Net corpus in the synonymous relationship called as Synsets. In the Word-Net, words are organized in the form of Synset. There are 207016 word-pairs presents in compressed form. Word-Net contain different type of semantic relationship like synonym, antonym, hyponyms and meronyms in noun, verb, adverb and adjective.  The Co-occurrence approach[ 2] formalizes the concept concerning information retrieval. In the word co-occurrence methodolgy, there is a word-list and for each word in the word list a meaningful word is connected.The query is retrieved by creating a vector. Ordering and context of this particular search query is fully unmeasured. So it is a major drawback of this methodolgy. In the lexical database methodology[ 3], similarity computation is done by the pre-defined Word-Net hierarchy, which is arranged in tree-like structure <ref type="bibr">[3]</ref>. Web search engine methodolgy compute the similarity but sometimes the word have opposite meaning which may occurred in the same webpage. This influenced the calculated similarity value adversly. This methodolgy is developed by Google-Similarity-Distane <ref type="bibr" target="#b0">[4]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Overview Of Different Similarity Measure Technique</head><p>The computation of similarity between two word pair i.e noun-noun pair/ verb-verb pair is done by using Word-Net ontology. Word-Net is developed by professor G. <ref type="bibr">A Miller[1]</ref> and it is managed by the laboratory of Conginitive science in princeton university. The ontology contain three databases noun, verb and adverb-adjective. In this ontology, words are organized in the form of Synsets.This paper is focussed on IS-A relationship between noun pair and verb pair. There are several hierarchy present in the Word-Net and all the hierarchy are subsumes in a common root node. Similarity approach are classified in different forms:</p><p>(a) Similarity calculation by distance based methodology (b) Similarity calculation by Information Content based methodology.</p><p>(c) Similarity calculation by Feature based methodogy.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1.">Similarity calculation by distance-based methodology</head><p>In this approach similarity is computed by measuring the distance between two words. This is also called Edge-Counting based methodology. Pair of words concerning Path Length is calculated for measuring the similarity among group of words. Similarity score measured by this approach is in discrete form, so there is applying normalization. Various path-based algorithms have been developed by Leacock-Chodorow <ref type="bibr" target="#b3">[7]</ref> and Wu and palmer <ref type="bibr" target="#b5">[9]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1.1.">Leacock-Chodorow Similarity Approach</head><p>The Approach of LCH <ref type="bibr" target="#b3">[7]</ref> is based on shortest path length. In this approach computation of similarity is done by finding the path length. It is the shortest path between noun-pair in the Word-Net IS-A relationship. The shortest path is defined as there are less no of intermediate node between two words. The shortest value retrieved by this is scaled up by the depth factor D where calculation of depth is done by the longest path from the root to leaf in the hierarchy of Word-Net. The calculation of similarity is done by.</p><formula xml:id="formula_0">𝑆𝑖𝑚𝑖𝑙𝑎𝑟𝑖𝑡𝑦 𝐿𝐶𝐻 = −𝑙𝑜𝑔 { (𝑚𝑖𝑛𝑖𝑚𝑢𝑚(𝑙𝑒𝑛𝑔𝑡ℎ(𝑤 1 , 𝑤 2 ))) 2 * 𝐷 } (1)</formula><p>where w1 denotes first word and w2 denotes second word. minimum(length(w1,w2)) denotes the minimum path length between word pair w1 and w2. Depth factor D is the maximum depth from root to leaf in the Word-Net.</p><p>LCH approach is easy. Computation of similarity between word pair is done by counting the number of links between word-pairs.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1.2.">Wu and Palmer Similarity Approach</head><p>The Wu and Palmer Similarity Approach <ref type="bibr" target="#b5">[9]</ref> focuses on path length among word-pair. This is based on most specific common predecessor node.It is called as Lowest Common Subsumer node(LCS).The Similarity between two words in a IS-A relationship in Word-net ontology is computed in this manner. This method perform well in verb ontology and different part of speeches where words are arranged in hierarichal structure. The calculation of similarity is done by:  </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.">Similarity calculation by Information Content</head><p>It deals with the similarity characteristic is calculated on the basis of information content. The content of Information concerning the word is computed by the frequency of the word in the Word-net ontology. The characteristic of frequency concerning the word is computed by the probability of occurrence of the word. Information content of the word is computed as:</p><formula xml:id="formula_1">Information Content(IC)= −𝑙𝑜𝑔(𝑝(𝑐)) (3)</formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.1.">Resnik Similarity Approach</head><p>Resnik <ref type="bibr" target="#b6">[10]</ref> similarity approach is based on the value of information content(IC) of the word.In this approach Similarity value calculation is done by how much information is shared between words w1 and w2. If the information shared between words are more than similarity value is high, otherwise low. The calculation of similarity is done by: SimilarityRES(w1,w2)= IC(lcs(w1,w2) <ref type="bibr" target="#b0">(4)</ref> This approach of computing similarity work on verb along with noun. The justification for this that both part of speech in structured in hierarichal manner <ref type="bibr" target="#b1">[5,</ref><ref type="bibr" target="#b2">6]</ref>.The value computed by resnik simlarity is always greater than 0.The Information Content value give better result than path-based approaches.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.2.">Similarity Approach of Jiang &amp; Conrath</head><p>Jiang and Conrath <ref type="bibr" target="#b7">[11]</ref> approach depends on the information content of the LCS between words and semantic distance between word pairs. The calculation of similarity is done by: 𝑫𝒊𝒔𝒕𝒂𝒏𝒄𝒆 𝑱𝒊𝒂𝒏𝒈&amp;𝑪𝒐𝒏𝒓𝒂𝒕𝒉 (𝒘 𝟏 , 𝒘 𝟐 ) = 𝑰𝑪𝒗𝒂𝒍𝒖𝒆(𝒘 𝟏 ) + 𝑰𝑪𝒗𝒂𝒍𝒖𝒆(𝒘 𝟐 ) − 𝟐 * 𝑰𝑪𝒗𝒂𝒍𝒖𝒆(𝒍𝒄𝒔(𝒘 𝟏 , 𝒘 𝟐 ))</p><p>(5)</p><formula xml:id="formula_2">𝑺𝒊𝒎𝒊𝒍𝒂𝒓𝒊𝒕𝒚 𝑱𝒊𝒂𝒏𝒈𝒂𝒏𝒅𝑪𝒐𝒏𝒓𝒂𝒕𝒉 = 𝟏 𝑫𝒊𝒔𝒕𝒂𝒏𝒄𝒆 𝑱𝒊𝒂𝒏𝒈&amp;𝑪𝒐𝒏𝒓𝒂𝒕𝒉 (𝒘 𝟏 , 𝒘 𝟐 )<label>(6)</label></formula><p>Where Distancejiang and conrath find the value of dissimilarity between words. The measure of values (low) indicate more similar words and high value indicate least similar words.</p><p>Jiang and Conrath approach is same as Resnik's approach. Similarity find by this approach is based on commonality between pair of words w1 and w2 and IC value of the words. There is special case need to handle when the value is 0.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.3.">Lin Similarity Approach</head><p>Lin similarity approach <ref type="bibr" target="#b8">[12]</ref> is based on the information that is shared by the word pairs to the summation of Information content value of the the word pairs. The concept of Lin's is based on the commonality in word-pairs to the information content value that described them completely.</p><p>𝑺𝒊𝒎 𝑳𝒊𝒏 = 𝟐 * 𝐈𝐂(𝐥𝐜𝐬(𝒘 𝟏 , 𝒘 𝟐 , )) * 𝐈𝐂(𝒘 𝟏 )+ 𝐈𝐂(𝒘 𝟐 ) <ref type="bibr" target="#b3">(7)</ref> Commonality means the information that is shared by the common ancestor node of the word w1 and w2 and the total amount of information described completely by the word w1 and w2. Lin computes the similarity between 0 and 1. 0 denotes the low similarity that mean two words are of different context and 1 denote the high similarity that mean two word are same or compared with itself. The similarity is calculated by: </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.3.">Similarity Computation by Feature based methodology</head><p>This similarity is based on features. Similarity value is high if the two words share the more common features and similarity value is low if the words have some unique features.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.3.1.">Tversky's Similarity Approach</head><p>Tversky's similarity approach <ref type="bibr" target="#b9">[13]</ref> is based on features. If the two word pairs w1 and w2 have more common features then similarity is high, and if the word pairs have indegenious features then the meaure of similarity values will be low. The Measure of Similarity between two words is based on how much the two words shared the common feature <ref type="bibr" target="#b4">[8]</ref>, the unique feature present in word w1 but not present in word w2 , and the unique feature present in word w2 but not present in word w1. The similarity is calculated by: </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Results &amp; Discussion</head><p>The result shows that on the implementation of the various edge counting methodology like LCH, Wu &amp; Palmer and Information content methodology like Resnik's and Jiang etc. The information content methodolgies gives better correlation than all the edge counting methodolgies.Comparison of all the similarity approach is shown in Fig <ref type="figure" target="#fig_5">6</ref> given below: </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Conclusion</head><p>Similarity between word-pairs is one of the emerging concept in the field of artificial intelligence, machine learning and genes priortization.Calculation of similarity between Word-pair is done by various approaches like distance based , information content based and feature based. All the approaches uses ontology of speific domain to find similarity.The Jiang and Conrath provides better result than other approaches. It is based on information and It has been seen that the similarity increases with the increase in the heirarchy of WordNet depth. So depth feature can be taken into account to find the similarity between words. The future task is to develop the approach which compute the simlarity between two pair of sentence</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">References</head><p>[1] Miller, George A., Richard Beckwith, Christiane Fellbaum, Derek Gross, and Katherine J.  </p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: A Fragment of WordNet taken from [3] for illustration</figDesc><graphic coords="2,72.00,281.43,445.69,270.85" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head></head><label></label><figDesc>B define as count of IS-A node between word w1 &amp; w2 to the most common ancestor node. C is defined as depth. It is computed from the word present in the Word-Net to the root in the Word-Net. The Approach of Wu and Palmer is based on the lowest common subsumer of two words. The value of similarity between words never becomes 0. In the Fig2.1. Comparison of similarity between words is determine by LCH and Wu &amp;Palmer based.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: Comparison of similarity between LCH ,Wu and palmer with human based judgement</figDesc><graphic coords="4,97.25,238.10,400.20,250.20" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Figure 3 :</head><label>3</label><figDesc>Figure 3: Comparison of similarity between Resnik's, Jiang &amp;Conrath and Lin with Human based Judgement's</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_4"><head>Figure 4 :</head><label>4</label><figDesc>Figure 4: Tversky Feature based Approach 𝑺𝒊𝒎𝒊𝒍𝒂𝒓𝒊𝒕𝒚 𝒕𝒗𝒆𝒓𝒔𝒌𝒚'𝒔 = 𝒂. 𝑮(𝒇(𝒘𝟏) ∩ 𝒇(𝒘𝟐)) − 𝒃𝑮(𝒇(𝒘𝟏) 𝒇(𝒘𝟐) − 𝒄𝑮(𝒇(𝒘𝟐) 𝒇(𝒘𝟏)</figDesc><graphic coords="6,86.20,524.16,435.65,130.80" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_5"><head>Figure 6 :</head><label>6</label><figDesc>Figure 6: Semantic similarity measure vs Correlation</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_6"><head></head><label></label><figDesc>Miller. "Introduction to WordNet: An on-line lexical database." International journal of lexicography 3, no. 4 (1990): 235-244. [2] Boyce, Bert R., Bert R. Boyce, Charles T. Meadow, Donald H. Kraft, Donald H. Kraft, and Charles T. Meadow. Text information retrieval systems. Elsevier, 2017. [3] Miller, George A. "WordNet: a lexical database for English." Communications of the ACM 38, no. 11 (1995): 39-41.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1</head><label>1</label><figDesc>Noun relationship in Word-Net (Corpus)</figDesc><table><row><cell>Relationship</cell><cell>Definition</cell><cell>Example</cell></row><row><cell>Hypernym</cell><cell>concept to subordinate</cell><cell>word1="breakfast"→word2="meal"</cell></row><row><cell></cell><cell>relationship</cell><cell></cell></row><row><cell>Hyponym</cell><cell>concept to sub-type</cell><cell>word1="plant"→word2="tree"</cell></row><row><cell></cell><cell>relationship</cell><cell></cell></row><row><cell>Part-of</cell><cell>part to whole relationship</cell><cell>word1="course"→word2="meal"</cell></row><row><cell>Has-part</cell><cell>whole to part relationship</cell><cell>word1="table"→word2="leg"</cell></row><row><cell>Antonym</cell><cell>opposite relationship</cell><cell>word1="leader"→word2="follower"</cell></row><row><cell>Hypernym</cell><cell>concept to subordinate</cell><cell>word1="breakfast"→word2="meal"</cell></row><row><cell></cell><cell>relationship</cell><cell></cell></row></table></figure>
		</body>
		<back>
			<div type="annex">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>𝑓(𝑤 2 ) 𝑓(𝑤 1 )</head><p>means the unique feature of w2 but not present in w1. f(w1)intersection f(w2) means feature similar in words.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.3.2.">The Similarity Approach of Piarro</head><p>Piarro approach <ref type="bibr" target="#b10">[14]</ref> focuses on the feature based approach developed by Tversky's . This approach explored the Tversky's feature based approach into information content(IC) domain. The function 𝐺(𝑥(𝑤 1 )𝑖𝑛𝑡𝑒𝑟𝑠𝑒𝑐𝑡𝑖𝑜𝑛𝑥(𝑤 2 ) is equivalent to IC(lcs(w1,w2)).</p><p>is equivalent to IC(w1)-IC(lcs(w1,w2)) ,</p><p>is equivalent to IC(w2)-IC(lcs(w1,w2)).</p><p>Where: ICvalue(lcs(w1,w2)) denotes Information content value of the subsumer. ICvalue(w1) denote Information-content value of the word w1. ICvalue(w2) denote the Information -content value of the word w2.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Comparison of Various Similarity Technique</head><p>On Comparison of two distance-based similarity approach i.e., Leacock and Chodorow and Wu and Palmer and three Information content-based approach similarity approach i.e., Resnik's and Lin etc. The result shows that similarity score focuses on the synonym of the words. Rubenstein <ref type="bibr" target="#b11">[15]</ref> selected 51 word-pair from human based judgement from set of 65 word-pairs. Rating of word-pair is given in the range from 0.0 to 4.0. The rating 0 means the words are semantically unrelated and rating 4.0 means words are highly related. Miller and Charles <ref type="bibr" target="#b12">[16]</ref> selected 30 set of word-pair from Rubenstein and Goodenough <ref type="bibr" target="#b11">[15]</ref> 65 set of word-pair. Miller's divide 30 word-pair into three sets i.e., 10 word-pair in each set. In the first set of 10 word-pair having rating between 3.0 to 4.0 are high similarity value words , next set of 10 word-pair having rating between 1.0 to 3.0 are intermediate similarity value words and last set of 10-word pair having rating between 0.0 to 1.0 are low similarity value words. In this paper for similarity calculation Miller and Charles 30 word-pair are taken for computation of similarity. In the  </p></div>			</div>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">The google similarity distance</title>
		<author>
			<persName><forename type="first">Rudi</forename><forename type="middle">L</forename><surname>Cilibrasi</surname></persName>
		</author>
		<author>
			<persName><surname>Paul</surname></persName>
		</author>
		<author>
			<persName><surname>Vitanyi</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on knowledge and data engineering</title>
		<imprint>
			<biblScope unit="volume">19</biblScope>
			<biblScope unit="issue">3</biblScope>
			<biblScope unit="page" from="370" to="383" />
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<monogr>
		<title level="m" type="main">Classification of Semantic Similarity Technique between Word Pairs using Word Net</title>
		<author>
			<persName><forename type="first">Atul</forename><surname>Gupta</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Krishan</forename><forename type="middle">Kumar</forename><surname>Goyal</surname></persName>
		</author>
		<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Computation of Verb Similarity</title>
		<author>
			<persName><forename type="first">Krishan</forename><surname>Goyal</surname></persName>
		</author>
		<author>
			<persName><surname>Kumar</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Design Engineering</title>
		<imprint>
			<biblScope unit="page" from="4127" to="4140" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Combining local context and WordNet similarity for word sense identification</title>
		<author>
			<persName><forename type="first">Claudia</forename><surname>Leacock</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Martin</forename><surname>Chodorow</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">WordNet: An electronic lexical database</title>
		<imprint>
			<biblScope unit="volume">49</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page" from="265" to="283" />
			<date type="published" when="1998">1998</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<monogr>
		<title level="m" type="main">Semantic similarity measure using information content approach with depth for similarity calculation</title>
		<author>
			<persName><forename type="first">Atul</forename><surname>Gupta</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Dharamveer</forename><surname>Kr</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Yadav</forename></persName>
		</author>
		<imprint>
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<monogr>
		<title level="m" type="main">Verb semantics and lexical selection</title>
		<author>
			<persName><forename type="first">Zhibiao</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Martha</forename><surname>Palmer</surname></persName>
		</author>
		<idno>arXiv preprint cmp- lg/9406033</idno>
		<imprint>
			<date type="published" when="1994">1994</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<monogr>
		<title level="m" type="main">Using information content to evaluate semantic similarity in a taxonomy</title>
		<author>
			<persName><forename type="first">Philip</forename><surname>Resnik</surname></persName>
		</author>
		<idno>arXiv preprint cmp-lg/9511007</idno>
		<imprint>
			<date type="published" when="1995">1995</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<monogr>
		<title level="m" type="main">Semantic similarity based on corpus statistics and lexical taxonomy</title>
		<author>
			<persName><forename type="first">Jay</forename><forename type="middle">J</forename><surname>Jiang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">David</forename><forename type="middle">W</forename><surname>Conrath</surname></persName>
		</author>
		<idno>arXiv preprint cmp-lg/9709008</idno>
		<imprint>
			<date type="published" when="1997">1997</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">An approach for measuring semantic similarity between words using multiple information sources</title>
		<author>
			<persName><forename type="first">Yuhua</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Zuhair</forename><forename type="middle">A</forename><surname>Bandar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">David</forename><surname>Mclean</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on knowledge and data engineering</title>
		<imprint>
			<biblScope unit="volume">15</biblScope>
			<biblScope unit="issue">4</biblScope>
			<biblScope unit="page" from="871" to="882" />
			<date type="published" when="2003">2003</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Features of similarity</title>
		<author>
			<persName><forename type="first">Amos</forename><surname>Tversky</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Psychological review</title>
		<imprint>
			<biblScope unit="volume">84</biblScope>
			<biblScope unit="issue">4</biblScope>
			<biblScope unit="page">327</biblScope>
			<date type="published" when="1977">1977</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">A semantic similarity metric combining features and intrinsic information content</title>
		<author>
			<persName><forename type="first">Giuseppe</forename><surname>Pirró</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Data &amp; Knowledge Engineering</title>
		<imprint>
			<biblScope unit="volume">68</biblScope>
			<biblScope unit="issue">11</biblScope>
			<biblScope unit="page" from="1289" to="1308" />
			<date type="published" when="2009">2009</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Contextual correlates of synonymy</title>
		<author>
			<persName><forename type="first">Herbert</forename><surname>Rubenstein</surname></persName>
		</author>
		<author>
			<persName><forename type="first">John</forename><forename type="middle">B</forename><surname>Goodenough</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Communications of the ACM</title>
		<imprint>
			<biblScope unit="volume">8</biblScope>
			<biblScope unit="issue">10</biblScope>
			<biblScope unit="page" from="627" to="633" />
			<date type="published" when="1965">1965</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Contextual correlates of semantic similarity</title>
		<author>
			<persName><forename type="first">George</forename><forename type="middle">A</forename><surname>Miller</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Walter</forename><forename type="middle">G</forename><surname>Charles</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Language and cognitive processes</title>
		<imprint>
			<biblScope unit="volume">6</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="1" to="2" />
			<date type="published" when="1991">1991</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
