<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Enriching very large ontologies using the WWW</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Eneko</forename><surname>Agirre</surname></persName>
						</author>
						<author>
							<persName><forename type="first">Olatz</forename><surname>Ansa</surname></persName>
						</author>
						<author>
							<persName><forename type="first">Eduard</forename><surname>Hovy</surname></persName>
						</author>
						<author>
							<persName><forename type="first">David</forename><surname>Martínez</surname></persName>
						</author>
						<title level="a" type="main">Enriching very large ontologies using the WWW</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">148EC59557D26408C19A65B776535FA2</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T12:27+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>This paper explores the possibility to exploit text on the world wide web in order to enrich the concepts in existing ontologies. First, a method to retrieve documents from the WWW related to a concept is described. These document collections are used 1) to construct topic signatures (lists of topically related words) for each concept in WordNet, and 2) to build hierarchical clusters of the concepts (the word senses) that lexicalize a given word. The overall goal is to overcome two shortcomings of WordNet: the lack of topical links among concepts, and the proliferation of senses. Topic signatures are validated on a word sense disambiguation task with good results, which are improved when the hierarchical clusters are used.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">INTRODUCTION</head><p>Knowledge acquisition is a long-standing problem in both Artificial Intelligence and Computational Linguistics. Semantic and world knowledge acquisition pose a problem with no simple answer. Huge efforts and investments have been made to build repositories with such knowledge (which we shall call ontologies for simplicity) but with unclear results, e.g. CYC <ref type="bibr" target="#b0">[1]</ref>, EDR <ref type="bibr" target="#b1">[2]</ref>, WordNet <ref type="bibr" target="#b2">[3]</ref>. WordNet, for instance, has been criticized for its lack of relations between topically related concepts, and the proliferation of word senses.</p><p>As an alternative to entirely hand-made repositories, automatic or semi-automatic means have been proposed for the last 30 years. On the one hand, shallow techniques are used to enrich existing ontologies <ref type="bibr" target="#b3">[4]</ref> or to induce hierarchies <ref type="bibr" target="#b4">[5]</ref>, usually analyzing large corpora of texts. On the other hand, deep natural language processing is called for to acquire knowledge from more specialized texts (dictionaries, encyclopedias or domain specific texts) <ref type="bibr" target="#b5">[6]</ref> <ref type="bibr" target="#b6">[7]</ref>. These research lines are complementary; deep understanding would provide specific relations among concepts, whereas shallow techniques could provide generic knowledge about the concepts.</p><p>This paper explores the possibility to exploit text on the world wide web in order to enrich WordNet. The first step consists on linking each concept in WordNet to relevant document collections in the web, which are further processed to overcome some of WordNet's shortcomings.</p><p>On the one hand, concepts are linked to topically related words. Topically related words form the topic signature for each concept in the hierarchy. As in <ref type="bibr" target="#b7">[8]</ref> <ref type="bibr" target="#b8">[9]</ref> we define a topic signature as a family of related terms {t, &lt;(w 1 ,s 1 )…(w i ,s i )…&gt;}, where t is the topic (i.e. the target concept) and each w i is a word associated with the topic, with strength s i . Topic signatures resemble relevancy signatures <ref type="bibr" target="#b9">[10]</ref>, but are not sentence-based, do not require parsing to construct, and are not suitable for use in information extraction. Topic signatures were originally developed for use in text summarization.</p><p>On the other hand, given a word, the concepts that lexicalize it (its word senses) are hierarchically clustered <ref type="bibr" target="#b10">[11]</ref>, thus tackling sense proliferation inWordNet.</p><p>Evaluation of automatically acquired semantic and world knowledge information is not an easy task. In this case we chose to perform task-oriented evaluation, via word sense disambiguation. That is, we used the topic signatures and hierarchical clusters to tag a given occurrence of a word with the intended concept. The benchmark corpus for evaluation is SemCor <ref type="bibr" target="#b11">[12]</ref>. Our aim is not to compete with other word sense disambiguation algorithms, but to test whether the acquired knowledge is valid.</p><p>This paper describes preliminary experiments. Several aspects could be improved and optimized but we chose to pursue the entire process first, in order to decide whether this approach is feasible and interesting. The resulting topical signatures and hierarchical clusters and their use on word sense disambiguation provide exciting perspectives.</p><p>The structure of the paper follows the same spirit: we first explain our method and experiments, and later review some alternatives, shortcomings and improvements. Section two reviews the ontology used and the benchmark corpus for word sense disambiguation. Next the method to build the topic signatures is presented, and a separate section shows the results on a word sense disambiguation task. The clustering method is presented alongside the associated word sense disambiguation results. Related work is discussed in the following section, and finally some conclusions are drawn and further work is outlined.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">BRIEF INTRODUCTION TO WORDNET AND SEMCOR</head><p>WordNet is an online lexicon based on psycholinguistic theories <ref type="bibr" target="#b2">[3]</ref>. It comprises nouns, verbs, adjectives and adverbs, organized in terms of their meanings around lexical-semantic relations, which include among others, synonymy and antonymy, hypernymy and hyponymy (similar to is-a links), meronymy and holonymy (similar to part-of links). Lexicalized concepts, represented as sets of synonyms called synsets, are the basic elements of WordNet. The version used in this work, WordNet 1.6, contains 121,962 words and 99,642 concepts. The noun boy, for instance, has 4 word senses, i.e. lexicalized concepts. The set of synonyms for each sense and the gloss is shown below:</p><p>1: male child, boy, childa youthful male person 2: boya friendly informal reference to a grown man 3: son, boya male human offspring 4: boyoffensive term for Black man Being one of the most commonly used semantic resources in natural language processing, some of its shortcomings are broadly acknowledged:</p><p>1. It lacks explicit links among semantic variant concepts with different part of speech; for instance paint-to paint or song-to sing are not related.</p><p>2. Topically related concepts are not explicitly related: there is no link between pairs like bat-baseball, fork-dinner, farmchicken, etc.</p><p>3. The proliferation of word sense distinctions in WordNet, which is difficult to justify and use in practical terms, since many of the distinctions are unclear. Line for instance has 32 word senses. This makes it very difficult to perform automatic word sense disambiguation.</p><p>This paper shows how to build lists of words that are topically related to a topic (a concept). These lists can be used to overcome the shortcomings just mentioned. In particular we show how to address the third issue, using the lists of words to cluster word senses according to the topic.</p><p>SemCor <ref type="bibr" target="#b11">[12]</ref> is a corpus in which word sense tags (which correspond to WordNet concepts) have been manually included for all open-class words in a 360,000-word subset of the Brown Corpus. We use SemCor to evaluate the topic signatures in a word sense disambiguation task. In order to choose a few nouns to perform our experiments, we focused on a random set of 20 nouns which occur at least 100 times in SemCor. The set comprises commonly used nouns like boy, child, action, accident, church, etc. These nouns are highly polysemous, with 6.3 senses on average.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">BUILDING TOPIC SIGNATURES FOR THE CONCEPTS IN WORDNET</head><p>In this work we want to collect for each concept in WordNet the words that appear most distinctively in texts related to it. That is, we aim at constructing lists of closely related words for each concept. For example, WordNet provides two possible word senses or concepts for the noun waiter: 1: waiter, server -a person whose occupation is to serve at table (as in a restaurant) 2: waiter -a person who waits or awaits For each of these concepts we would expect to obtain two lists with words like the following:</p><p>1: restaurant, menu, waitress, dinner, lunch, counter, etc. 2: hospital, station, airport, boyfriend, girlfriend, cigarette, etc. The strategy to build such lists is the following (cf. Figure <ref type="figure" target="#fig_0">1</ref>). We first exploit the information in WordNet to build queries, which are used to search in the Internet those texts related to the given word sense. We organize the texts in collections, one collection per word sense. For each collection we extract the words and their frequencies, and compare them with the data in the other collections. The words that have a distinctive frequency for one of the collections are collected in a list, which constitutes the topic signature for each word sense.</p><p>The steps are further explained below.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">Building the queries</head><p>The original goal is to retrieve from the web all documents related to an ontology concept. If we assume that such documents have to contain the words that lexicalize the concept, the task can be reduced to classifying all documents where a given word occurs into a number of collections of documents, one collection per word sense. If a document cannot be classified, it would be assigned to an additional collection. The goal as phrased above is unattainable, because of the huge amount of documents involved. Most of words get millions of hits: boy would involve retrieving 2,325,355 documents, church 6,243,775, etc. Perhaps in the future a more ambitious approach could be tried, but at present we cannot aim at classifying those enormous collections. Instead, we construct queries, one per concept, which are fed to a search engine. Each query will retrieve the documents related to that concept.</p><p>The queries are constructed using the information in the ontology. In the case of WordNet each concept can include the following data: words that lexicalize the concept (synonyms), a gloss and examples, hypernyms, hyponyms, meronyms, holonyms and attributes. Altogether a wealth of related words is available, which we shall call cuewords. If a document contains a high number of such cuewords around the target word, we can conclude that the target word corresponds to the target concept. The cuewords are used to build a query which is fed into a search engine, retrieving the collection of related documents.</p><p>As we try to constrain the retrieved documents to the 'purest' documents, we build the queries for each word sense trying to discard documents that could belong to more than one sense. For instance, the query for word x in word sense i (being j,k other word senses for x) is constructed as follows:</p><p>(x AND (cueword 1,i OR cueword 2,i ...) AND NOT (cueword 1,j OR cueword 2,j ... OR cueword 1,k OR cueword 2,k ...) where cueword l,m stands for the cueword l of word sense m. This boolean query searches for documents that contain the target word together with one of the cuewords of the target concept, but do not contain any of the cuewords of the remaining concepts. If a cueword appears in the information relative to more than one sense, it is discarded.</p><p>Deciding which of the cuewords to use, and when, is not an easy task. For instance, nouns in the definition are preferable to the other parts of speech, monosemous cuewords are more valuable than polysemous ones, synonyms provide stronger evidence than meronyms, other concepts in the hierarchy can also be used, etc. After some preliminary tests, we decided to experiment with all available information: synonyms, hypernyms, hyponyms, coordinate sisters, meronyms, holonyms and nouns in the  </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2">Search the internet</head><p>Once the queries are constructed we can use a number of different search engines. We started to use just the first 100 documents from a list of search engines. This could bias the documents, and some could be retrieved repeatedly. Therefore, unlike <ref type="bibr" target="#b12">[13]</ref>, we decided to use only one search engine, the most comprehensive search engine at the time, AltaVista [14]. AltaVista allows complex queries which were not possible in some of the other web search engines.</p><p>The number of documents retrieved for the 20 words amounts to the tens of thousands, taking more that one gigabyte of disk space once compressed, and 9 days of constant internet access. For instance, it took 3 hours and a half to retrieve the 1,217 documents for the four senses of boy, which took 100 megabytes once compressed.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3">Build topic signatures</head><p>The document collections retrieved in step 3.2 are used to build the topic signatures. The documents are processed in order to extract the words in the text. We did not perform any normalization; the words are collected as they stand. The words are counted and a vector is formed with all words and their frequencies in the document collection. We thus obtain one vector for each collection, that is, one vector for each word sense of the target word.</p><p>In order to measure which words appear distinctively in one collection in respect to the others, a signature function was selected based on previous experiments <ref type="bibr" target="#b13">[15]</ref> <ref type="bibr" target="#b12">[13]</ref>. We needed a function that would give high values for terms that appear more frequently than expected in a given collection. The signature function that we used is χ 2 , which we will define next.</p><p>The vector vf i contains all the words and their frequencies in the document collection i, and is constituted by pairs (word j , freq i,j ), that is, one word j and the frequency of the word j in the document collection i. We want to construct another vector vx i with pairs (word j , w i,j ) where w i,j is the χ 2 value for the word j in the document collection i (cf. Equation <ref type="formula">1</ref>). Equation <ref type="formula" target="#formula_0">2</ref>defines m i,j , the expected mean of word j in document i.</p><p>When computing the χ 2 values, the frequencies in the target document collection are compared with the rest of the document collection, which we call the contrast set. In this case the contrast set is formed by the other word senses. Excerpts from the signatures for boy are shown in Table <ref type="table" target="#tab_2">2</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">APPLY SIGNATURES FOR WORD SENSE DISAMBIGUATION</head><p>The goal of this experiment is to evaluate the automatically constructed topic signatures, not to compete against other word sense disambiguation algorithms. If topic signatures yield good results in word sense disambiguation, it would mean that topic signatures have correct information, and that they are useful for word sense disambiguation. Given the following sentence from SemCor, a word sense disambiguation algorithm should decide that the intended meaning for waiter is that of a restaurant employee:</p><p>"There was a brief interruption while one of O'Banion's men jerked out both his guns and threatened to shoot a waiter who was pestering him for a tip." Word sense disambiguation is a very active research area (cf. <ref type="bibr" target="#b14">[16]</ref> for a good review of the state of the art). Present word sense disambiguation systems use a variety of information sources <ref type="bibr" target="#b15">[17]</ref> which play an important role, such us collocations, selectional restrictions, topic and domain information, co-occurrence relations, etc. Topic signatures constitute one source of evidence, but do not replace the others. Therefore, we do not expect impressive results.</p><p>The word sense disambiguation algorithm is straightforward. Given an occurrence of the target word in the text we collect the words in its context, and for each word sense we retrieve the χ 2 values for the context words in the corresponding topic signature.  </p><formula xml:id="formula_0">m i,j = Σ i freq i,j Σ j freq i,j Σ i,j freq i,j<label>(2)</label></formula><p>For each word sense we add these χ 2 values, and then select the word sense with the highest value. Different context sizes have been tested in the literature, and large windows have proved to be useful for topical word sense disambiguation <ref type="bibr" target="#b16">[18]</ref>. We chose a window of 100 words around the target.</p><p>In order to compare our results, we computed a number of baselines. First of all choosing the sense at random (ran). We also constructed lists of related words using WordNet, in order to compare their performance with that of the signatures: the list of synonyms (Syn), these plus the content words in the definitions (S+def), and these plus the hyponyms, hypernyms and meronyms (S+all). The algorithm to use these lists is the same as for the topic signatures.</p><p>Table <ref type="table" target="#tab_3">3</ref> shows the results for the selected nouns. The number of senses attested in SemCor 3 (#s) and the number of occurrences of the word in SemCor (#occ) are also presented. The results are given as precision, that is, the number of successful tags divided by the total number of occurrences. A precision of one would mean that all occurrences of the word are correctly tagged.</p><p>The results show that, the precision of the signature-based word sense disambiguation (Sign column) is well above the precision for random selection (a few exceptions are in bold), and, that, overall, it outperforms the other WordNet-based lists of words (the winner for each word is in bold). This proves that topic signatures managed to learn topic information that was not originally present in WordNet. This information is overly correct, but in some cases introduces noise and the performance degrades even below the random baseline (e.g. action, hour).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">CLUSTERING WORD SENSES</head><p>In principle we could try to cluster all the concepts in WordNet, comparing their topic signatures, but instead we experimented with clustering just the concepts that belong to a given word (its word senses). As we mentioned in Section 2, WordNet makes very fine distinctions between word senses, and suffers excessive word sense proliferation.</p><p>For many practical applications we can ignore some of the sense distinctions. For instance, all of the senses for boy are persons. 3 Some word senses never occur in SemCor. We did not take those senses into account.</p><p>Two of the senses refer to young boys while two of them refer to grown males. 'Boy as a young person' would tend to appear in a certain kind of documents, while 'boy as a grown man' in others, and 'boy as a colored person' in yet other documents. In this work, as in <ref type="bibr" target="#b13">[15]</ref>[13], we tried to compare the overlap between the signatures by simply counting shared words, but this did not yield interesting results. Instead we used binary hierarchical clustering directly on the retrieved documents <ref type="bibr" target="#b10">[11]</ref>. We experimented with various distance metrics and clustering methods but the results did not vary substantially: slink <ref type="bibr" target="#b17">[19]</ref>, clink <ref type="bibr" target="#b18">[20]</ref>, median, and Ward's method <ref type="bibr" target="#b19">[21]</ref>. Some of the resulting hierarchies were analyzed by hand and they were coherent according to our own intuitions. For instance Figure <ref type="figure">2</ref> shows that the young and offspring senses of boy (nodes 1 and 3) are the closest (similarity of 0.65), while the informal (node 2) and colored (node 4) senses are further apart. The contexts for the colored sense are the least similar to the others (0.46).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.1">Evaluation of word sense clusters on a word sense disambiguation task</head><p>Hand evaluation of the hierarchies is a difficult task, and very hard to define <ref type="bibr" target="#b10">[11]</ref>. As before, we preferred to evaluate them on a word sense disambiguation task. We devised two methods to apply the hierarchies and topic signatures to word sense disambiguation:</p><p>1. Use the original topic signatures. In each branch of the hierarchy we combine all the signatures for the word senses in the branch, and choose the highest ranking branch. For instance, when disambiguating boy, we first choose between boy4 and the rest: boy1, boy2, boy3 (cf. Figure <ref type="figure">2</ref>). Given a occurrence, the evidence for boy1, boy2, boy3 is combined, and compared to the evidence for boy4. The winning branch is chosen. If boy4 is discarded, then the combined evidence for boy1, boy3 is compared to that of boy2. If boy2 gets more evidence, that is the chosen sense.</p><p>2. Build new topic signatures for the existing clusters. The document collections for all the word senses in the branch are merged and new χ 2 values are computed for each cluster in the hierarchy. For instance, at the first level we would have a topic The word sense disambiguation algorithm can be applied at different levels of granularity, similar to decision trees. At the first level it chooses to differentiate between boy4 and the rest, at the second level among boy4, boy2 and boy1-3, and at the third level it disambiguates the finest-grained senses.</p><p>Instead of evaluating the set of all nouns, we focused on three nouns: boy, cell and church. The results are shown in Table <ref type="table" target="#tab_4">4</ref>. The second column shows the number of senses. The signature results for the original sense distinctions (cf. Table <ref type="table" target="#tab_3">3</ref>) are shown in the second column. The results for the signature and hierarchy combination are shown according to the sense-distinctions: the fine column shows the results using the hierarchy for the finest sense distinctions, the medium column corresponds to the medium sized clusters, and the coarse level corresponds to the coarsest clusters, i.e., all senses clustered in two groups. For each level, three results are given: the random baseline, the results using the original topic signatures and the hierarchy, and the results with the new topic signatures computed over the clusters (best results for in bold). The results show that the information contained in the hierarchy helps improve the precision obtained without hierarchies, even at the fine level. For coarser sense distinctions it exceeds 0.90 precision. Regarding the way to apply the hierarchy, the results are not conclusive. Further experiments would be needed to show whether it is useful or not to compute new topic signatures for each cluster.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6">DISCUSSION AND COMPARISON WITH RELATED WORK</head><p>The work here presented involves different areas of research. We will focus on the method to build topic signatures, the method to cluster the concepts and how the document collection for each word sense is constructed.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.1">Building topic signatures</head><p>Topic signatures were an extension of relevancy signatures <ref type="bibr" target="#b9">[10]</ref> developed for text summarization <ref type="bibr" target="#b13">[15]</ref>. To identify topics in documents, <ref type="bibr" target="#b13">[15]</ref> constructed topic signatures from 16,137 documents classified into 32 topics of interest. His topic signature construction method is similar to ours, except that he used tf.idf for term weighting. In subsequent work, Hovy and Junk <ref type="bibr" target="#b12">[13]</ref> explored several alternative weighting schemes in a topic identification task, finding that χ 2 provided better results than tf.idf or tf, and that specific combinations of χ 2 and Latent Semantic Analysis provided even better results on clean training data. Lin and Hovy <ref type="bibr" target="#b8">[9]</ref> use a likelihood ratio from maximum likelihood estimates that achieves even better performance on clean data. However, their experiments with text extracted from the web proved somewhat disappointing, like the ones reported here.</p><p>In general, documents retrieved from the web introduce a certain amount of noise into signatures. The results are still useful to identify the word sense of the target words, as our results show, but a hand evaluation of them is rather worrying. We concluded that the cause of the poor quality does not come from the procedure to build the signatures, but rather from the quality of the documents retrieved (c.f. Section 6.3).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.2">Concept Clustering</head><p>Traditional clustering techniques <ref type="bibr" target="#b10">[11]</ref> are difficult to apply to concepts in ontologies. The reason is that the usual clustering methods are based on statistical word co-occurrence data, and not on concept co-occurrence data (which is not available at present). The method presented in this paper uses the fact that concepts are linked to document collections. Usual document clustering techniques are applied to document collections, effectively clustering the associated concepts. This clustering method tackles the word sense proliferation WordNet.</p><p>The evaluation and validation of the word sense clusters is difficult <ref type="bibr" target="#b10">[11]</ref>. We chose to evaluate the performance of the clusters in a word sense disambiguation task, showing that the clusters are useful to improve the results; enhanced precision for the fine-level sense distinctions, and over 90% precision for the coarse level.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.3">Searching the internet for concepts</head><p>The core component of the method explored in this paper is the technique to link documents in the web to concepts in an ontology. Recently, some methods have been explored to automatically retrieve examples for concepts from large corpora and the internet. Leacock et al. <ref type="bibr" target="#b20">[22]</ref> use a strategy based on the monosemous relatives of WordNet concepts to retrieve examples from a 30 million word corpus. As their goal is to find 100 examples for each word sense of a given word, they prefer close relatives such us synonyms or hyponym collocations that contain the target hyponym. If enough examples are not found, they also use other hyponyms, sisters and hypernyms. The examples were used to train a supervised word sense disambiguation algorithm with very good results, but no provision was made to enrich WordNet with them. The main shortcoming of this strategy is that limiting the search to monosemous relatives, only 65% of the concepts under study could get training examples.</p><p>Mihalcea and Mondovan <ref type="bibr" target="#b21">[23]</ref> present a similar work which tries to improve the previous method. When a monosemous synonym for a given concept is not found, additional information from the definition of the concept is used, in the form of defining phrases constructed after parsing and processing the definition. The whole internet is used as a corpus, using a search engine to retrieve the examples. Four procedures are defined to query the search engine in order: use monosemous synonyms, use the defining phrases, use synonyms with the AND operator and words from the defining phrase with the NEAR operator, and lastly, use synonyms and words from the defining phrases with the AND operator. The procedures are sorted by preference, and one procedure is only applied if the previous one fails to retrieve any examples. 20 words totaling 120 senses were chosen, and an average of 670 examples could be retrieved for each word sense. The top 10 examples for each word sense were hand-checked and 91% were found correct.</p><p>Both these methods focus on obtaining training examples. In contrast, our method aims at getting documents related to the concept. This allows us to be less constraining; the more documents the better, because that allows to found more distinctively co-occurring terms. That is why we chose to use all close relatives for a given concept, in contrast to <ref type="bibr" target="#b20">[22]</ref> which only focuses on monosemous relatives, and <ref type="bibr" target="#b21">[23]</ref>, which uses synonyms and a different strategy to process the gloss. Another difference is that our method forbids the cuewords of the rest of the senses.</p><p>We have found that searching the web is the weakest point of our method. The quality and performance of the topic signatures and clusters depends on the quality and number of the retrieved documents, and our query strategy is not entirely satisfactory. On the one hand some kind of balance is needed. For some querying strategies some word senses do not get any document, and with other strategies too many and less relevant documents are retrieved. On the other hand the web is not a balanced corpus (e.g. the sexual content in the topic signatures for boy). Besides, many documents are short indexes or cover pages, with little text on them. In this sense, the query construction has to be improved and some filtering techniques should be devised.</p><p>Other important consideration about searching the internet is that technical features have to be taken in consideration. For instance, our system had some timeout parameters, meaning that the retrieval delay of the documents (caused by the hour, workload, localization of server, etc.) could affect the results.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="7">CONCLUSIONS AND FURTHER RESEARCH</head><p>We have introduced an automatic method to enrich very large ontologies, e.g. WordNet, that uses the huge amount of documents in the world wide web. The core of our method is a technique to link document collections from the web to concepts, which allows to alleviate some of the main problems acknowledged in WordNet; lack of relations between topically related concepts, and the proliferation of word senses. We show in practice that the document collections can be used 1) to create topic signatures (lists of words that are topically related to the concept) for each WordNet concept, and, 2) given a word, to cluster the concepts that lexicalize it (its word senses), thus tackling sense proliferation. In order to validate the topic signatures and word sense clusters, we demonstrate that they contain information which is useful in a word sense disambiguation task. This work combines several techniques, and we chose to pursue the whole method from start to end. This strategy left much room for improvement in all steps. Both signature construction and clustering seem to be satisfactory, as other work has also shown. In particular, nice clean signatures are obtained when constructing topic signatures from topically organized documents. On the contrary, topic signatures extracted from the web seem to be dirtier.</p><p>We think that, in this work, the main obstacle to get clean signatures comes from the method to link concepts and relevant documents from the web. The causes are basically two. First, the difficulty to retrieve documents relevant to one and only one concept. The query construction has to be improved and carefully fine-tuned to overcome this problem. Second, the wild and noisy nature of the texts in the web, with its high bias towards some topics, high number of not really textual documents e.g., indexes., etc. Some filtering techniques have to be applied in order to get documents with less bias and more content.</p><p>Cleaner topic signatures open the avenue for interesting ontology enhancements, as they provide concepts with rich topical information. For instance, similarity between topic signatures could be used to find out topically related concepts, the clustering strategy could be extended to all concepts rather that just the concepts that lexicalize the same word, etc. Besides, word sense disambiguation methods could profit from these richer ontologies, and improve word sense disambiguation performance.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 .</head><label>1</label><figDesc>Figure 1. Overall design.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1</head><label>1</label><figDesc>shows part of the information available for sense 1 of boy.The query for sense 1 of boy would include the above cuewords plus the negation for the cuewords of the other senses. An excerpt of the query: (boy AND ('altar boy' OR 'ball boy' OR ...OR 'male person)</figDesc><table><row><cell>Target word</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell></cell><cell></cell><cell>sense1 + information</cell><cell></cell><cell>Query1</cell><cell></cell><cell>Document</cell><cell></cell><cell>Topic</cell></row><row><cell>Word Net</cell><cell>Look-up</cell><cell>sense2 + information ... senseN + information</cell><cell>Build queries</cell><cell>Query2 ... QueryN</cell><cell>Query WWW</cell><cell>collectionN Docum Docum collection2 collection1</cell><cell>Build Signatures</cell><cell>signatureN Docume signature1 Docume signature2</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 1 .</head><label>1</label><figDesc>Information for sense 1 of boy.</figDesc><table><row><cell>synonyms</cell><cell>male child, child</cell></row><row><cell>gloss</cell><cell>a youthful male person</cell></row><row><cell cols="2">hypernyms male, male person</cell></row><row><cell>hyponyms</cell><cell>altar boy, ball boy, bat boy, cub, lad, laddie, sonny,</cell></row><row><cell></cell><cell>sonny boy, boy scout, farm boy, plowboy, ...</cell></row><row><cell>coordinate</cell><cell>chap, fellow, lad, gent, fella, blighter, cuss, foster</cell></row><row><cell>systers</cell><cell>brother, male child, boy, child, man, adult male, ...</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 2 .</head><label>2</label><figDesc>Top words in signatures for three senses of boy.</figDesc><table><row><cell>Boy1</cell><cell>Boy2</cell><cell>Boy3</cell></row><row><cell>(child 9854)</cell><cell>(gay 7474)</cell><cell>(human 5023)</cell></row><row><cell>(Child 5979)</cell><cell>(reference 5154)</cell><cell>(son 4898)</cell></row><row><cell>(person 4671)</cell><cell>(tpd-results 3930)</cell><cell>(Human 3055)</cell></row><row><cell>(anything.com 3702)</cell><cell>(sec 3917)</cell><cell>(Soup 1852)</cell></row><row><cell>(Opportunities 1808)</cell><cell>(gay 2906)</cell><cell>(interactive 1842)</cell></row><row><cell>(Insurance 1796)</cell><cell>(Xena 1604)</cell><cell>(hyperinstrument 1841)</cell></row><row><cell>(children 1458)</cell><cell>(male 1370)</cell><cell>(Son 1564)</cell></row><row><cell>(Girl 1236)</cell><cell>(ADD 1304)</cell><cell>(clips 1007)</cell></row><row><cell>(Person 1093)</cell><cell>(storing 1297)</cell><cell>(father 918)</cell></row><row><cell>(Careguide 918)</cell><cell>(photos 1203)</cell><cell>(man-child 689)</cell></row><row><cell>(Spend 839)</cell><cell>(merr 1077)</cell><cell>(measure 681 )</cell></row><row><cell>(Wash 821)</cell><cell>(accept 1071)</cell><cell>(focus 555)</cell></row><row><cell>(enriching 774)</cell><cell>(PNorsen 1056)</cell><cell>(research 532)</cell></row><row><cell>(prizes 708)</cell><cell>(software 1021)</cell><cell>(show 461)</cell></row><row><cell>(Scouts 683)</cell><cell>(adult 983)</cell><cell>(Teller 456)</cell></row><row><cell>(Guides 631)</cell><cell>(penny 943)</cell><cell>(Yo-Yo 455)</cell></row><row><cell>(Helps 614)</cell><cell>(PAGE 849)</cell><cell>(modalities 450)</cell></row><row><cell>(Christmas 525)</cell><cell>(Sex 835)</cell><cell>(performers 450)</cell></row><row><cell>(male 523)</cell><cell>(Internet 725)</cell><cell>(senses 450)</cell></row><row><cell>(address 504)</cell><cell>(studs 692)</cell><cell>(magicians 448)</cell></row><row><cell>(paid 472)</cell><cell>(porno 675)</cell><cell>(percussion 439)</cell></row><row><cell>(age 470)</cell><cell>(naked 616)</cell><cell>(mother 437)</cell></row><row><cell>(mother 468)</cell><cell>(erotic 611)</cell><cell>(entertainment 391)</cell></row><row><cell>...up to 6.4 Mbytes</cell><cell>...up to 4.4 Mbytes</cell><cell>... up to 4.7 Mbytes</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 3 .</head><label>3</label><figDesc>Word sense disambiguation results. Hierarchy for the word senses of boy signature for boy4 and another for the merged collections of boy1, boy2 and boy3. At the second level we would have a topic signature for boy2 and another for boy1, boy3.</figDesc><table><row><cell>Word</cell><cell cols="3">#s #occ Ran Syn S+def S+all Sign</cell></row><row><cell>Accident</cell><cell>2</cell><cell>12 0.50 0.00 0.56</cell><cell>0.71 0.50</cell></row><row><cell>Action</cell><cell>8</cell><cell>130 0.12 0.00 0.05</cell><cell>0.29 0.02</cell></row><row><cell>Age</cell><cell>3</cell><cell>104 0.33 0.01 0.04</cell><cell>0.03 0.60</cell></row><row><cell>Amount</cell><cell>4</cell><cell>103 0.25 0.22 0.27</cell><cell>0.30 0.50</cell></row><row><cell>Band</cell><cell>7</cell><cell>21 0.14 0.11 0.13</cell><cell>0.28 0.25</cell></row><row><cell>Boy</cell><cell>4</cell><cell>169 0.25 0.45 0.37</cell><cell>0.59 0.66</cell></row><row><cell>Cell</cell><cell>3</cell><cell>116 0.33 0.00 0.37</cell><cell>0.36 0.59</cell></row><row><cell>Child</cell><cell>2</cell><cell>206 0.50 0.37 0.47</cell><cell>0.43 0.29</cell></row><row><cell>Church</cell><cell>3</cell><cell>128 0.33 0.28 0.50</cell><cell>0.46 0.45</cell></row><row><cell>Difference</cell><cell>5</cell><cell>112 0.20 0.02 0.28</cell><cell>0.35 0.17</cell></row><row><cell>Door</cell><cell>4</cell><cell>138 0.25 0.05 0.24</cell><cell>0.26 0.04</cell></row><row><cell>Experience</cell><cell>3</cell><cell>125 0.33 0.22 0.42</cell><cell>0.35 0.42</cell></row><row><cell>Fact</cell><cell>4</cell><cell>124 0.25 0.02 0.48</cell><cell>0.58 0.82</cell></row><row><cell>Family</cell><cell>6</cell><cell>135 0.17 0.12 0.18</cell><cell>0.15 0.36</cell></row><row><cell>Girl</cell><cell>5</cell><cell>152 0.20 0.34 0.21</cell><cell>0.33 0.25</cell></row><row><cell>History</cell><cell>5</cell><cell>104 0.20 0.06 0.16</cell><cell>0.17 0.18</cell></row><row><cell>Hour</cell><cell>2</cell><cell>110 0.50 0.21 0.63</cell><cell>0.38 0.40</cell></row><row><cell>Information</cell><cell>3</cell><cell>146 0.33 0.00 0.12</cell><cell>0.64 0.66</cell></row><row><cell>Plant</cell><cell>2</cell><cell>99 0.50 0.30 0.42</cell><cell>0.45 0.82</cell></row><row><cell>World</cell><cell>8</cell><cell>210 0.12 0.09 0.18</cell><cell>0.19 0.34</cell></row><row><cell>Overall</cell><cell cols="2">83 2444 0.28 0.16 0.30</cell><cell>0.36 0.41</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_4"><head>Table 4 :</head><label>4</label><figDesc>Results using hierarchies and word sense clusters</figDesc><table><row><cell></cell><cell></cell><cell>Sign</cell><cell cols="2">Signature &amp; Hierarchy</cell></row><row><cell cols="2">Word #</cell><cell cols="3">Fine Orig Rand Orig New Rand Orig New Rand Orig New Medium Coarse</cell></row><row><cell>Boy</cell><cell cols="4">4 0.66 0.25 0.68 0.38 0.33 0.83 0.67 0.50 0.99 0.99</cell></row><row><cell>Cell</cell><cell cols="3">3 0.59 0.33 0.62 0.52 -</cell><cell>-</cell><cell>-0.50 0.52 0.96</cell></row><row><cell cols="4">Church 3 0.45 0.33 0.48 0.54 -</cell><cell>-</cell><cell>-0.50 0.77 0.90</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">IxA NLP group. University of the Basque Country. 649 pk.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="20" xml:id="foot_1">.080 Donostia. Spain. Email: eneko@si.ehu.es, jipanoso@si.ehu.es. jibmaird@si.ehu.es 2 USC Information Sciences Institute, 4676 Admiralty Way, Marina del Rey, CA 90292-6695, USA. Email: hovy@isi.edu.</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>ACKNOWLEDGEMENTS</head><p>We would like to thank the referees for their fruitful comments. Part of the work was done while Eneko Agirre was visiting ISI, funded by the Basque Government.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">CYC: A Large-Scale Investment in Knowledge Infrastructure</title>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">B</forename><surname>Lenat</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Communications of the ACM</title>
		<imprint>
			<biblScope unit="volume">38</biblScope>
			<biblScope unit="issue">11</biblScope>
			<date type="published" when="1995">1995</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">The EDR Electronic Dictionary</title>
		<author>
			<persName><forename type="first">T</forename><surname>Yokoi</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Communications of the ACM</title>
		<imprint>
			<biblScope unit="volume">38</biblScope>
			<biblScope unit="issue">11</biblScope>
			<date type="published" when="1995">1995</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<monogr>
		<title level="m" type="main">Wordnet: An Electronic Lexical Database</title>
		<author>
			<persName><forename type="first">C</forename><surname>Fellbaum</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1998">1998</date>
			<publisher>MIT Press</publisher>
			<pubPlace>Cambridge</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Customizing a Lexicon to Better Suit a Computational Task</title>
		<author>
			<persName><forename type="first">M</forename><surname>Hearst</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Schütze</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of the Workshop on Extracting Lexical Knowledge</title>
				<meeting>of the Workshop on Extracting Lexical Knowledge</meeting>
		<imprint>
			<date type="published" when="1993">1993</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Automatic construction of a hypernym-labeled noun hierarchy from text</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">A</forename><surname>Caraballo</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of the Conference of the Association for Computational Linguistics</title>
				<meeting>of the Conference of the Association for Computational Linguistics</meeting>
		<imprint>
			<date type="published" when="1999">1999</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<monogr>
		<author>
			<persName><forename type="first">Y</forename><surname>Wilks</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><forename type="middle">M</forename><surname>Slator</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Guthrie</surname></persName>
		</author>
		<title level="m">Electric Words: Dictionaries, Computers, and Meanings</title>
				<meeting><address><addrLine>Cambridge</addrLine></address></meeting>
		<imprint>
			<publisher>MIT Press</publisher>
			<date type="published" when="1996">1996</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">WordNet 2 -A Morphologically and Semantically Enhanced Resource</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">M</forename><surname>Harabagiu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">A</forename><surname>Miller</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">I</forename><surname>Moldovan</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of the SIGLEX Workshop</title>
				<meeting>of the SIGLEX Workshop</meeting>
		<imprint>
			<date type="published" when="1999">1999</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Automated Text Summarization in SUMMARIST</title>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">H</forename><surname>Hovy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C.-Y</forename><surname>Lin</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Advances in Automatic Text Summarization</title>
				<editor>
			<persName><forename type="first">M</forename><surname>Maybury</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">I</forename><surname>Mani</surname></persName>
		</editor>
		<meeting><address><addrLine>Cambridge</addrLine></address></meeting>
		<imprint>
			<publisher>MIT Press</publisher>
			<date type="published" when="1999">1999</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">The Automated Acquisition of Topic Signatures for Text Summarization</title>
		<author>
			<persName><forename type="first">C.-Y</forename><surname>Lin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">H</forename><surname>Hovy</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of the COLING Conference</title>
				<meeting>of the COLING Conference<address><addrLine>Strasbourg, France</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2000-08">2000. August, 2000</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">An Empirical Study of Automated Dictionary Construction for Information Extraction in Three Domains</title>
		<author>
			<persName><forename type="first">E</forename><surname>Riloff</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Artificial Intelligence</title>
		<imprint>
			<biblScope unit="volume">85</biblScope>
			<date type="published" when="1996">1996</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Clustering Algorithms</title>
		<author>
			<persName><forename type="first">E</forename><surname>Rasmussen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Information Retrieval: Data Structures and Algorithms</title>
				<editor>
			<persName><forename type="first">W</forename><forename type="middle">B</forename><surname>Frakes</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">R</forename><surname>Baeza-Yates</surname></persName>
		</editor>
		<meeting><address><addrLine>London</addrLine></address></meeting>
		<imprint>
			<publisher>Prentice Hall</publisher>
			<date type="published" when="1992">1992</date>
			<biblScope unit="page" from="419" to="442" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">A Semantic Concordance</title>
		<author>
			<persName><forename type="first">G</forename><surname>Miller</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Leacock</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Tengi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Bunker</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of ARPA Workshop on Human Language Technology</title>
				<meeting>of ARPA Workshop on Human Language Technology</meeting>
		<imprint>
			<date type="published" when="1993">1993</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<monogr>
		<title level="m" type="main">Using Topic Signatures to Enrich the SENSUS Ontology</title>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">H</forename><surname>Hovy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Junk</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1998">1998</date>
		</imprint>
	</monogr>
	<note>In prep</note>
</biblStruct>

<biblStruct xml:id="b13">
	<monogr>
		<title level="m" type="main">Robust Automated Topic Identification</title>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">Y</forename><surname>Lin</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1997">1997</date>
		</imprint>
		<respStmt>
			<orgName>University of Southern California</orgName>
		</respStmt>
	</monogr>
	<note type="report_type">PhD thesis</note>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Introduction to the Special Issue on Word Sense Disambiguation: The State of the Art</title>
		<author>
			<persName><forename type="first">N</forename><surname>Ide</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Veronis</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Computational Linguistics</title>
		<imprint>
			<biblScope unit="volume">24</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="1" to="40" />
			<date type="published" when="1998">1998</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Using Multiple Knowledge Sources for Word Sense Discrimination</title>
		<author>
			<persName><forename type="first">S</forename><surname>Mcroy</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Computational Linguistics</title>
		<imprint>
			<biblScope unit="volume">18</biblScope>
			<biblScope unit="issue">1</biblScope>
			<date type="published" when="1992">1992</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">A method for Disambiguating Word Senses in a Large Corpus</title>
		<author>
			<persName><forename type="first">W</forename><forename type="middle">A</forename><surname>Gale</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><forename type="middle">W</forename><surname>Church</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Yarowsky</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Computer and the Humanities</title>
		<imprint>
			<biblScope unit="volume">26</biblScope>
			<biblScope unit="page" from="415" to="439" />
			<date type="published" when="1993">1993</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">SLINK: an Optimally efficient algorithm for the Single-Link Cluster Method</title>
		<author>
			<persName><forename type="first">R</forename><surname>Sibson</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Computer Journal</title>
		<imprint>
			<biblScope unit="volume">16</biblScope>
			<biblScope unit="page" from="30" to="34" />
			<date type="published" when="1973">1973</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">An Efficient Algorithm for a Complete Link Method</title>
		<author>
			<persName><forename type="first">D</forename><surname>Defays</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Computer Journal</title>
		<imprint>
			<biblScope unit="volume">20</biblScope>
			<biblScope unit="page" from="364" to="366" />
			<date type="published" when="1977">1977</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">Application of an Hierarchical Grouping Procedure to a Problem of Grouping Profiles</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">H</forename><surname>Ward</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">E</forename><surname>Jr</surname></persName>
		</author>
		<author>
			<persName><surname>Hook</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Educational and Psychological Measurement</title>
		<imprint>
			<biblScope unit="volume">23</biblScope>
			<biblScope unit="page" from="69" to="81" />
			<date type="published" when="1963">1963</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">Using Corpus Statistics and WordNet Relations for Sense Identification</title>
		<author>
			<persName><forename type="first">C</forename><surname>Leacock</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Chodorow</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">A</forename><surname>Miller</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Computational Linguistics</title>
		<imprint>
			<biblScope unit="volume">24</biblScope>
			<biblScope unit="issue">2</biblScope>
			<date type="published" when="1998">1998</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">An Automatic Method for Generating Sense Tagged Corpora</title>
		<author>
			<persName><forename type="first">R</forename><surname>Mihalcea</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">I</forename><surname>Moldovan</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of the Conference of the American Association of Artificial Intelligence</title>
				<meeting>of the Conference of the American Association of Artificial Intelligence</meeting>
		<imprint>
			<date type="published" when="1999">1999</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
