<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Designing clustering methods for ontology building: The Mo&apos;K workbench</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Gilles</forename><surname>Bisson</surname></persName>
						</author>
						<author>
							<persName><forename type="first">Claire</forename><surname>Nédellec</surname></persName>
						</author>
						<author>
							<persName><forename type="first">Dolores</forename><surname>Cañamero</surname></persName>
						</author>
						<title level="a" type="main">Designing clustering methods for ontology building: The Mo&apos;K workbench</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">68B6CF91B304DECDCF1B912E5D3FCCCE</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T12:26+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>This paper describes Mo'K, a configurable workbench that supports the development of conceptual clustering methods for ontology building. Mo'K is intended to assist ontology developers in the exploratory process of defining the most suitable learning methods for a given task.</p><p>To do so, it provides facilities for evaluation, comparison, characterization and elaboration of conceptual clustering methods. Also, the model underlying Mo'K permits a finegrained definition of similarity measures and class construction operators, easing the tasks of method instantiation and configuration. This paper presents some experimental results that illustrate the suitability of the model to help characterize and assess the performance of different methods that learn semantic classes from parsed corpora.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">INTRODUCTION</head><p>In this paper we propose a workbench that supports the development of conceptual clustering methods for the (semi-) automatic construction of ontologies of a conceptual hierarchy type from parsed corpora. The elaboration of any clustering method involves the definition of two main elements-a distance metrics and a classification algorithm. In the context of conceptual hierarchy formation, the Natural Language Processing (NLP) community has investigated the notion of distance to elaborate the semantic classes underlying hierarchies. Classification algorithms have been broadly studied within the Machine Learning and Data Analysis communities.</p><p>Different tools have been developed for the automatic or semi-automatic acquisition of semantic classes from "near" terms. The notion of semantic proximity is based upon distance among terms, defined as a function of the degree of similarity of the contexts. Descriptions of term contexts (the learning examples) and of the regularities to be sought vary in different approaches. Contexts can be purely graphic-words cooccurring within a window-as in the case of <ref type="bibr" target="#b0">[1]</ref>, <ref type="bibr" target="#b3">[4]</ref>; in some cases, the window can cover the whole document (see e.g. <ref type="bibr" target="#b20">[21]</ref>). Contexts can also be syntactic, as in the approaches that we have taken into account to develop our model, e.g. <ref type="bibr" target="#b12">[13]</ref>, <ref type="bibr" target="#b13">[14]</ref>, <ref type="bibr" target="#b10">[11]</ref>, <ref type="bibr" target="#b19">[20]</ref>, <ref type="bibr" target="#b4">[5]</ref>, <ref type="bibr" target="#b25">[26]</ref>, <ref type="bibr" target="#b6">[7]</ref>. However, the selection of a suitable distance for a given corpus and task is still an open problem that has not received much attention so far <ref type="bibr" target="#b24">[25]</ref>. In most cases, the criteria proposed to support this choice rely on the evaluation of the application task for which learning takes place, as described for instance in <ref type="bibr" target="#b11">[12]</ref>. Evaluation criteria proposed to assess learning results are purely quantitative, and comparative analyses of these criteria are rare <ref type="bibr" target="#b4">[5]</ref>. A proper characterization of the effects that different methods have on the learning results would provide methodological guidelines to help the designer select the most suitable method for a given corpus and task, or to provide support to create a new one.</p><p>This observation also applies to classification algorithms. No meth odology or tool has been proposed to support the elaboration of conceptual clustering algorithms that build taskspecific ontologies. Work on conceptual clustering (e.g., <ref type="bibr" target="#b18">[19]</ref>, <ref type="bibr" target="#b7">[8]</ref>, <ref type="bibr" target="#b8">[9]</ref>, <ref type="bibr" target="#b1">[2]</ref>, <ref type="bibr" target="#b0">[1]</ref>, <ref type="bibr" target="#b25">[26]</ref>) has not been extensively applied to the problem of learning from corpora. One must however acknowledge that the application of conceptual clustering techniques to this domain is not straightforward, as existing algorithms must be previously adapted. As in the case of distances, the elaboration and selection of a suitable algorithm for a given corpus and task requires the development of new methodological guidelines and tools.</p><p>As a first step toward this goal we propose Mo'K, a configurable workbench to support the comparison, evaluation, and elaboration of methods to learn conceptual hierarchies. The conceptual clustering model underlying Mo'K permits a finegrained definition of the components of distances and of class construction operators, easing the tasks of method instantiation and configuration. The model is extended with a set of variables that permit to characterize features specific to the elaboration of learning corpora, such as pruning, stop-lists, etc. The workbench also includes evaluation criteria to assess learning results obtained for different parameter configurations. We finally present some experimental results that illustrate the suitability of the model to help characterize different methods and assess their performance. These results concern only class formation, not classification algorithms.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">FRAMEWORK</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1">Learning semantic classes</head><p>In the context of learning semantic classes, learning from syntactic contexts exploits syntactic relations among words to derive semantic relations, following Harris' hypothesis <ref type="bibr" target="#b14">[15]</ref>. According to this hypothesis, the study of syntactic regularities within a specialized corpus permits to identify syntactic schemata made out of combinations of word classes reflecting specific domain knowledge. The fact of using specialized corpora eases the learning task, given that we have to deal with a limited vocabulary with reduced polysemy, and limited syntactic variability.</p><p>In syntactic approaches, learning results can be of different types, depending on the method employed. They can be distances that reflect the degree of similarity among terms <ref type="bibr" target="#b12">[13]</ref>, <ref type="bibr" target="#b25">[26]</ref>, <ref type="bibr" target="#b21">[22]</ref>, distance-based term classes elaborated with the help of nearest-neighbor methods <ref type="bibr" target="#b10">[11]</ref>, <ref type="bibr" target="#b13">[14]</ref>, degrees of membership in term classes <ref type="bibr" target="#b23">[24]</ref>, class hierarchies formed by conceptual clustering <ref type="bibr" target="#b19">[20]</ref>, or predicative schemata that use concepts to constraint selection <ref type="bibr" target="#b0">[1]</ref>, <ref type="bibr" target="#b9">[10]</ref>, <ref type="bibr" target="#b6">[7]</ref>. The notion of distance is fundamental in all cases, as it allows to calculate the degree of proximity between two objects-terms in this case-as a function of the degree of similarity between the syntactic contexts in which they appear. Classes built by aggregation of near terms can afterwards be used for different applications, such as syntactic disambiguation <ref type="bibr" target="#b22">[23]</ref>, <ref type="bibr" target="#b23">[24]</ref> or document retrieval <ref type="bibr" target="#b10">[11]</ref>. Distances are however calculated using the same similarity notion in all cases, and our model relies on these studies regardless of the application task.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2">Conceptual Clustering</head><p>In our case, ontologies are organized as multiple hierarchies that form an acyclic graph where nodes are term categories described by intention, and li nks represent inclusion, seen in this case as a generality relation. Learning though hierarchical classification of a set of objects can be performed in two main ways: top-down, by incremental specialization of classes, and bottom-up, by incremental generalization. We have adopted a bottom-up approach due to its smaller algorithmic complexity, and its understandability to the user in view of an interactive validation task. In this article we focus on the elements needed to build and evaluate the basic classes of this graph, i.e. criteria for building the initial corpus, distances, and evaluation criteria to asses results. We do not address the generic class construction algorithm. With respect to this latter, let us just mention that the application of hierarchical (conceptual <ref type="bibr" target="#b18">[19]</ref> or numerical <ref type="bibr" target="#b5">[6]</ref>) clustering algorithms to our problem is not straightforward, given that we must build acyclic graphs with few abstraction levels, rather than deep and strict hierarchies.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">THE MO'K MODEL AND WORKBENCH</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">Representation of examples and results</head><p>Following the standard practice, we use binary grammatical relations as syntactic contexts. Examples are therefore represented by triplets &lt;Head -Grammatical Relation -Modifier head&gt;, where &lt;Modifier head&gt; is the object that must be classified and &lt;Head -Grammatical Relation&gt; represents the attribute. The number of occurrences of a triplet in a corpus characterizes the attribute for an example. For instance, if we are interested in verbal attachments, the following two sentences:</p><p>• This causes a decrease in <ref type="bibr">[…]</ref>.</p><p>• This high rate results from an increase in <ref type="bibr">[…]</ref>.</p><p>allow to generate two triplets, &lt;cause Dobj decrease &gt; (29), and &lt;Result from(Adj) increase &gt; (2), both presenting the structure &lt;Verb -grammatical relation-Head noun&gt; (total number of occurrences of these triads in the corpus). In the remaining of the paper, we will designate by Action the tuple &lt;Verbgrammatical relation&gt;. Actions &lt;Cause Dobj&gt; and &lt;Result from(Adj)&gt; can be regarded as objects, and nouns &lt;Decrease&gt; and &lt;Increase&gt; can be considered as attributes with values 29 and 2, respectively. In bottom-up clustering, couples of near objects or of objects and classes are incrementally grouped in order to form hierarchies or graphs of object classes. The standard in NLP is to use object classes (Figure <ref type="figure">1</ref>.a) for the application task. In our experiments, we depart from this practice to compare learned classes, as we are interested in an extensional representation; we therefore use classes formed by the union of attributes of near objects (Figure <ref type="figure">1</ref>.b). Let us take an example. If the objects &lt;Cause Dobj&gt; and &lt;Result from(Adj)&gt; are selected to form a class, their attribute sets are merged. Let us suppose that &lt;Cause Dobj&gt; is described by the nouns {decrease, increase, modification, loss, etc.}, and &lt;Result from(Adj)&gt; by {decrease, increase, composition, evolution, etc.}. The noun class learned will include nouns shared by both objects (in bold), and also the complementary terms (in italics); therefore, four new triplets are induced. We will then use the "attribute class" strategy, as th is way the "leaf" classes that we will later evaluate will be larger than those formed using the "object class" strategy. We will not further develop here the differences between these two viewpoints, intension and extension, since this topic is out of the scope of the paper. Let us however insist on the fact that the selection of one or the other has major effects on the learning results.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2">Corpus parameters</head><p>The parameters used to form a learning corpus in Mo'K include, among others, selection of learning examples, level of pruning, and "cleaning" of the corpus. Let us examine the first two.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.1">Selection of learning examples</head><p>One of the goals of our model is to allow the user to compare learning results as a function of the grammatical relations selected as input. Objects and syntactic contexts used in classification vary in different approaches-i.e. verbs or nouns which are considered as similar on the grounds of their shared verbal or nominal contexts, where nouns can be verb complement heads (arguments <ref type="bibr" target="#b13">[14]</ref>, or adjuncts <ref type="bibr" target="#b6">[7]</ref>), noun complements <ref type="bibr" target="#b10">[11]</ref> or all of them <ref type="bibr" target="#b4">[5]</ref>, <ref type="bibr" target="#b12">[13]</ref>. None of these approaches proposes a comparative study of results based on grammatical relations chosen in the initial corpus. Our model easily allows to specify these relations. Experiments concerning verbal relations reported below (Section 4.2) illustrate this and show significant differences among results depending on the nature of objects and attributes (whether they are nouns or verbs), and on the type of corpus.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.2">Pruning</head><p>A second parameter, taken into account by most existing methods and included in our model, concerns corpus pruning as a function of the number of occurrences of an element. Pruning removes occurrences that are too infrequent and therefore would cause noise, as well as those which are too frequent and do not provide any information regarding the link between an object and an attribute. The other side of the coin is that infrequent but important cases can be removed. Our model also allows to specify the minimum number of examples characterizing an attribute and the minimal number of attributes for an example and, for each of these constraints, the minimal total number of occurrences of the triplets being considered. The experiments reported in Section 4.3 show that the level of pruning has a major impact on the results of learning, and that the optimal level depends on the corpus.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3">Distance Modeling</head><p>Our goal is not to cover all the possible methods that can be used to measure similarity between examples. On the contrary, our approach focuses on methods with very precise features:</p><p>• They take syntactic analysis as input;</p><p>• They do not take into account external resources (e.g. ontologies such as WordNet <ref type="bibr" target="#b17">[18]</ref>); • They are based on a comparison of the distribution profiles of the attributes describing the couples of object to classify. Different methods have been proposed in the NLP literature with in this framework-among others <ref type="bibr" target="#b13">[14]</ref>, <ref type="bibr" target="#b10">[11]</ref> [13], <ref type="bibr" target="#b4">[5]</ref> and <ref type="bibr" target="#b6">[7]</ref>. We have developed a generic model of these methods and implemented it in Mo'K with the aim of elaborating a comparison and evaluation methodology for them. In order to come up with a generic implementation, we have identified the steps shared by all these methods, as we will see below. Mo'K is thus a workbench that implements a set of instantiable generic methods using an object-oriented representation, as opposed to the idea of a library of methods. This approach is made possible by the fact that similarity measures can in general be regarded as a comparison of the "distribution profiles" of couples of examples. This way, two objects will be considered as neighbors if the relative occurrence frequencies of each of their attributes (i.e. of the syntactic contexts) are close. Learning examples taken into account in our model can be represented by means of a contingency table. Depending on the representation hypothesis adopted, rows (examples) and columns (attributes) of the table represent different things. For example, in the experiments reported in Section 4.2 they first represent actions and nouns, respectively, and nouns and actions later on. In any case, a table cell contains the number of occurrences of an attribute for a given example. This table is obviously very sparse, as examples are generally described by a small number of attributes (see Figure <ref type="figure" target="#fig_1">2</ref>).</p><p>In practice, computation of similarity can be decomposed into two major st eps-weighting and similarity computation.</p><p>• The weighting phase changes every raw value of cooccurrences appearing in the contingency table by a coefficient, often normalized, which can be regarded as a weight or measure of the significance of the fact of examples and attributes co-occurring in the corpus. Its computation can entail two steps-the initialization of the weights of examples and attributes, usually according to their number of occurrences, and the calculation of a normalized weight of the relevance of each attribute for each example. Technically, this weighting phase is implemented in Mo'K using the 5 functions described (in pseudo-code) in Table <ref type="table" target="#tab_0">1</ref>.</p><p>• The similarity computation phase builds a similarity matrix between couples of examples. Similarity increases as a function of the number of shared attributes, but the way in which similarity between these distributions is calculated varies in the different approaches. In Mo'K this phase is implemented by a single function. We thus see that, by means of 6 functions and using a few lines of code, it is possible to implement most similarity measures that follow a schema based on comparison of pairs of distribution profiles. Let us no te that we do not make more specific hypotheses concerning the formal properties of measures-they can be similarities or dissimilarities, symmetrical or asymmetrical, and computed information can be of any type. This approach thus favors the comparison of existing methods, but also the elaboration of variants of these methods and even the creation of new ones. Once integrated in Mo'K, a method can access all the test and conceptual clustering resources of the system.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Name of the step Method</head><p>Initialization of the weight of each example E: W(E) Initialization of the weight of each attribute A: W(A) </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.4">Distance evaluation</head><p>Even though our goal is the construction of hierarchies, it is interesting to evaluate the relevance of a distance metrics with respect to more simple tasks and to analyze its behavior as a function of the application domain and of the parameters of elaboration of the learning corpus. Mo'K offers different means of evaluation based on the first N couples of examples built by binary aggregation, i.e. the first N couples of examples with highest scores in the similarity matrix.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.4.1">Measure of recall</head><p>As already mentioned in Section 3, the elaboration of a class gives rise to the induction of new triplets not observed in the initial corpus. Therefore, the evaluation process follows the classical schema of dividing the corpus in two partitions-learning and test. The former is used to build the similarity matrix according to the measure to be evaluated. The latter allows to measure the coverage rates of classes, i.e. their ability to recognize the triplets in the test set. We have adopted this evaluation task for two reasons. First, it corresponds to the elementary step in every process of bottom-up hierarchical clustering. Second, from a NLP perspective, it conforms to a disambiguation task.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.4.2">Measure of precision</head><p>Despite its interest, the coverage measure only allows to evaluate the recall rate associated with the set N of selected classes. However, precision-a measure of the ability to avoid erroneous recognition of negative examples-is an equally important property of the metrics. In the end, a similarity measure that tends to over-generalize and describe object couples using a large number of attributes would reach high coverage rates, but produce classes that lack in meaning and precision. As we will see in the experiments reported in next section, it is interesting to measure other criteria in order to assess the relevance of a similarity measure-for example, the induction rate measuring the ratio between the number of induced triplets and the total number of triplets learned.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">EXPERIMENTS AND RESULTS</head><p>The experiments reported here aim to illustrate Mo'K's parameterization possibilities and the impact that different parameter settings have on the learning results. These experiments make thus a case for the use of generic platforms to perform a systematic exploratory analysis in order to obtain sensible results in a given domain (corpus).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1">Training corpora</head><p>We have conducted experiments on two different French corpora-one contains cooking recipes gathered over the world wide web; the other, Agrovoc, contains scientific abstracts in the agricultural domain, and has been assembled by the INIST (Institut de l'Information Scientifique et Technique) of the CNRS. We have chosen these two corpora as they differ in generality and amount of technical terms, but are still close enough to allow for meaningful comparison of results. Both corpora have been analyzed using the same shallow parser. Only verbal relations of the form &lt;Verb -Grammatical Relation -Noun&gt; have been considered in our experiments. The output of syntactic parsing is highly noisy (between 30% and 50% mistakes) due to several factors such as grammatical and spelling mistakes, typos, and accentuation errors in the case of the cooking corpus. In Agrovoc, noise is mostly produced by the high number of technical terms mistaken by verbs, and of embedded noun complements which are erroneously attached to verbs. In Agrovoc, only 300 verbs are found versus 18,828 nouns. This is due to the fact that only part of the corpus has been considered-those triplets with a verb that appears in a list of verbs giving rise to nominalizations in the corpus. In the cooking corpus we find 1,181 verbs for 3,300 nouns, i.e. the ratio between nouns and verbs is, in average, divided by a factor of 20 with respect to Agrovoc. This reflects a higher specialization in this corpus. Finally, Agrovoc is three times as big as the cooking corpus (117,156 triplets for 168,287 occurrences in total).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2">Selection of example representation</head><p>The first experiment illustrates the importance of the choice of the object to be clustered and of the attribute which characterizes it.We have compared the classes learned by considering actions as objects and head nouns as attributes (denoted as action-based representation), with those learned using nouns as objects and actions as attributes (denoted as noun-based representation). The comparison (see Table <ref type="table" target="#tab_2">2</ref>) has been performed on both corpora, the other parameters remaining unchanged. We have applied Asium's distance. Pruning criteria are light-no minimum number of triplet occurrences, the minimum number of examples characterizing a given attribute has been set to 2, and the minimum number of attributes for an example to 3. We will further comment this setting in Section 4.3. For each corpus, the first 25 learned classes are evaluated. Coverage is measured on the test set, which comprises 20% of the whole corpus, and on the artificial set, which contains 50,000 triplets randomly generated (see Section 3.3). Each test has been repeated four times.</p><p>The first experiment has been conducted on Agrovoc. The classes learned in the action-based representation are twice as large as those learned in the noun-based representation. However, induction rates (number of induced triplets divided by the number of triplets learned by rote) are very similar (40% compared to 38%) and so are precision rates (45% in both cases). The experiments on the cooking corpus have built classes of similar size in both representations. Induction rate is slightly higher for the noun-based representation-38% compared to 34%. However, recall on the test set is better in the action-based representation, (12% and 9.1%, respectively). Induced triplets are thus more useful in the case of action-based representation, even though they are less numerous. Moreover, the precision measured by the rate of negative artificial examples covered by learned examples is much better for the action-based representation (32% compared to 52%). Precision and recall rates are thus better in this representation, although the rate of induced triplets is smaller than in the noun-based representation. In any case, all rates are much better than the ones computed for the Agrovoc experiments. A closer examination of the best pairs of actions and nouns confirms the idea that overgeneralization is less of a problem here than in Agrovoc. Noun pairs are more precise (e.g., [fridge-freezer], [olive oil-oil]) and described by more technical actions. In the same way, the best pairs of actions, such as [absorb Dobj , evaporate Dobj ], are characterized by nouns (in this case [vinegar, water, wine, excess, etc.]) which are significantly more specific than in Agrovoc. The smaller variability of the cooking corpus explains these observations, showing that the larger size of Agrovoc does not improve the meaningfulness of the regularities observed.</p><p>This experiment thus shows that the choice of a representation can have a major impact on the learning results. It is therefore advisable to select a suitable representation before addressing a new domain.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.3">Pruning parameters</head><p>To illustrate the importance of pruning, two pruning settings have been applied to both corpora and compared. In both cases we have used Asium's distance, objects are actions and attributes are nouns. The first setting is the one described in the previous section. In the second one, we have set the minimum number of occurrences for an example (triplet) to 2, in order to remove triplets occurring only once in the corpus, as they may represent noise. We have also augmented the values of the minimal number of examples that must characterize an attribute (from 2 to 3), and of the minimal number of attributes per object (from 3 to 5, i.e. in this version the objects being compared appear in at least 5 different syntactic contexts). As it can be inferred from the histogram in Fig. <ref type="figure" target="#fig_1">2</ref>, this setting excludes 80% of the corpus, versus 70% in the first setting. We can thus hope for a more reliable classification, to the risk of removing so many examples that coverage (recall) is drastically affected. The experiments show that, for the cooking corpus, induction rate (32%) and coverage (11.2%) on the whole test set are nearly unaffected. On the contrary, the rate of artificial triplets covered scales by a factor of 3 with respect to the previous rate; we think that this significant increase indicates that this pruning rate increases the rate of erroneously induced examples. In Agrovoc, induction rate decreases by a factor of three and coverage rate by a factor of two, whereas the rate of artificial triplets covered scales by a factor superior to 2. Recall is therefore much strongly affected by pruning than in the cooking corpus. In both cases, this new pruning setting gives rise to a decrease in performance. However, this conclusion drawn from the evaluation of basic classes must be tempered in the case of hierarchy formation. In this case, the more constraining version of pruning allows to eliminate many non-significant classes that result from the presence of closely similar actions described by a small number of attributes. It seems clear that, in a process of hierarchical clustering, this type of class would cause problems, as it would alter some groupings. Therefore, the type of pruning to be applied partly depends on the task to be performed.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.4">Comparison of methods</head><p>This last experiment illustrates some aspects of the use of Mo'K to compare results obtained with different distances. Among the methods that we have tested with Mo'K (such as those proposed by Dagan et al ., Hindle, Grefenstette, and Grishman et al . among the best known in the literature, as well as other distances proposed by the authors, e.g. Asium, and Greedy) we have chosen to compare here the distances used in Asium <ref type="bibr" target="#b6">[7]</ref>, the one proposed by Dagan <ref type="bibr" target="#b4">[5]</ref>, and Greedy<ref type="foot" target="#foot_2">3</ref> . The reason for th is choice relies on the fact that these three methods have shown good performance when compared with others, while presenting rather different behaviors. They are thus a representative sample of the subset of methods that we have modeled as characterized in Section 3.3. These methods have been applied to the corpus of cooking recipes. Learning parameters are the same as those described in Section 4.2, objects are actions and attributes are names. In addition, we have tested the influence on learning of the number of disjointed classes on which recall performance is evaluated. To do this, we have varied this number (in abscissa of the diagrams below) between 10 and 200. These diagrams show, respectively, the recall rate of each method on the test set (Figure <ref type="figure">4a</ref>), and the efficiency of classes rate (Figure <ref type="figure">4b</ref>). The latter is assessed by the ratio between the number of triplets learned (by rote and induced), and the number of triplets effectively used in the recall test. As we can see in the first diagram, the coverage rate of the three methods grows as expected according to the number of classes considered, Dagan's method yielding the best results. On the contrary, if we pay attention to the efficiency of classes, Asium takes better advantage of the triplets learned. Looking at both diagrams, we can conclude that these methods have different behaviors. Dagan's gives rise to more general classes (more triplets are learned, but the number of useless ones is higher). Asium constructs more specific classes (fewer triplets learned, but more of them are useful). We can take a closer look at the behavior of these methods and study the quality of induction in terms of the rate of induced triplets which are effectively used in the recall test (Figure <ref type="figure" target="#fig_3">5</ref>). While the previous conclusion is confirmed for Asium and Dagan's methods, we can also note that Greedy and Dagan's methods have the same behavior along this criterion. In fact, it seems that Dagan's method is able to induce more useful triplets than Greedy, whereas this latter tends to learn by rote a more representative sample subset of the corpus.</p><p>The tests performed on the artificial examples confirm these results. Therefore, it seems that the classes learned by Dagan's method and, to a lesser extent, by Greedy are less robust and present lower precision rates for of the learning parameters chosen. We must emphasize that these conclusions only apply to the recipe corpus. For Agrovoc, results are considerably different. These experiments therefore show the importance of going through an exploratory process in order to come up with the most suitable methods and representation.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">CONCLUSIONS AND PROSPECTS</head><p>Mo'K is a configurable workbench that supports the development of conceptual clustering methods for specific ontology building. The learning model proposed here takes parsed corpora as input. No additional (terminological or semantic) knowledge is used for labeling the input, guiding learning, or validating the learning results. Preliminary experiments showed that the quality of learning decreases with the generality of the corpus. This makes somehow unrealistic the use of general ontologies for guiding such learning, as they seem too incomplete and polysemic to allow for efficient learning in specific domains. For example, <ref type="bibr" target="#b15">[16]</ref> points out that 40% of the words in canonical form in the titles and abstracts of the Communications of the ACM are not included in the LDOCE (Longman Dictionary of Contemporary English). This problem posed in the case of learning specific ontologies obviously does no t apply in the case of guiding learning of general semantic classes, as shown in the abundant literature on the topic (see e.g. <ref type="bibr" target="#b22">[23]</ref>, <ref type="bibr" target="#b21">[22]</ref>, <ref type="bibr" target="#b23">[24]</ref>, <ref type="bibr" target="#b16">[17]</ref>). It would however be highly valuable to take advantage of existing ontologies to improve the quality of learning. We consider that this can be achieved in two main ways. First, learning could be improved by the use of specific terminologies, dictionaries and nomenclature, such as SNOMED International in the medical domain <ref type="bibr" target="#b2">[3]</ref>. Second, some methodological guidelines would be needed to integrate specialized learned ontologies into more general ontologies such as WordNet <ref type="bibr" target="#b17">[18]</ref>.</p><p>Although we have focused on a disambiguation-based task, other validation tasks could be integrated in Mo'K, such as query extension and information extraction. Learning specialized ontologies of high quality for these tasks will allow the development of applications in technical and rapidly evolving domains, in which manual acquisition is too costly. In this sense, we have started exploring information extraction from molecular biology abstracts.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 .Figure 1 .</head><label>11</label><figDesc>Figure 1.a. Object classes (= Nouns).</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 2 .</head><label>2</label><figDesc>Figure 2. Example distribution per number of attribute</figDesc><graphic coords="5,92.00,392.00,150.00,95.00" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Figure 4a&amp;b .</head><label>4a&amp;b</label><figDesc>Figure 4a&amp;b. Recall rate and Class efficiency</figDesc><graphic coords="5,308.00,195.00,118.00,78.00" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Figure 5 .</head><label>5</label><figDesc>Figure 5. Quality of induction.</figDesc><graphic coords="5,369.00,488.00,116.00,78.00" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1 .</head><label>1</label><figDesc>Functions implementing the weighting phase in Mo'K.</figDesc><table><row><cell></cell><cell>Init_Weight_Example</cell></row><row><cell></cell><cell>Init_Weight_Attribute</cell></row><row><cell>For each example E</cell><cell></cell></row><row><cell>For each attribute A of the example</cell><cell></cell></row><row><cell>Calculate W(A) in the context of E</cell><cell>Eval_Weight_Example</cell></row><row><cell>Update global W(E)</cell><cell>Eval_Weight_Attribute</cell></row><row><cell>For each attribute A of the example</cell><cell></cell></row><row><cell>Normalization of the W(A) by W(E)</cell><cell>Init_Similarity</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 2 .</head><label>2</label><figDesc>Precision rate represents the rate of negative examples among learned examples which cover artificial examples. The recall measured by the coverage rate of induced triplets on the test set is slightly better for the noun-based representation (5.3% compared to 4.7%) but remains quite low. This can be explained by the level of generality of what is learned: the best couples of learned nouns involved very general terms such as [technique-method], [influence-effect], in contrast with the numerous technical words of the corpus. Most of the actions characterizing these nouns concern general verbs such as [to present], [to observe], and [to report]. This is confirmed by looking at the best pairs of actions such as &lt;to study Dobj&gt; -&lt;to analyze Dobj&gt;. This explains why learned classes are of rather poor quality in both representations. Experimentation about example representation</figDesc><table><row><cell>Corpus</cell><cell>Learning</cell><cell>% Induced tripl./</cell><cell>Recall (test</cell><cell>Precision</cell></row><row><cell></cell><cell>object</cell><cell>learned tripl.</cell><cell>set)</cell><cell></cell></row><row><cell>Agrovoc</cell><cell>Action</cell><cell>40 %</cell><cell>4.7 %</cell><cell>45 %</cell></row><row><cell></cell><cell>Nom</cell><cell>38 %</cell><cell>5.3 %</cell><cell>45 %</cell></row><row><cell>Cooking</cell><cell>Action</cell><cell>34 %</cell><cell>12 %</cell><cell>32 %</cell></row><row><cell></cell><cell>Nom</cell><cell>38 %</cell><cell>9.1 %</cell><cell>52 %</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">HELIX Project, INRIA Rhône-Alpes, ZIRST, 655 Avenue de l'Europe, F-38330 Montbonnot, email: Gilles.Bisson@imag.fr</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_1">Inference and Learning Group, LRI, Bât. 490, CNRS UMR 8623 &amp; Université de Paris-Sud, F-91405 Orsay Cedex, email: {cn, lola}@lri.fr</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_2">Greedy has a simple behavior: it is based on a measure, inspired from the χ 2 , which favors the selection of pairs of examples described by a large number of attributes (the method is named after this feature).</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>ACKNOWLEDGEMENTS</head><p>We are grateful to INIST-CNRS for providing the Agrovoc corpus. This research is partly funded by the French Ministry of Industry under RNRT project Astuxe.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">An empirical symbolic approach to natural language processing</title>
		<author>
			<persName><forename type="first">R</forename><surname>Basili</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">T</forename><surname>Pazienza</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Velardi</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Artificial Intelligence Journal</title>
		<imprint>
			<biblScope unit="volume">85</biblScope>
			<biblScope unit="page" from="59" to="99" />
			<date type="published" when="1996">1996</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Conceptual Clustering in a First Order Logic Representation</title>
		<author>
			<persName><forename type="first">G</forename><surname>Bisson</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of 10th European Conference on Artificial Intelligence (ECAI&apos;92)</title>
				<meeting>10th European Conference on Artificial Intelligence (ECAI&apos;92)<address><addrLine>Vienna</addrLine></address></meeting>
		<imprint>
			<date type="published" when="1992">1992</date>
			<biblScope unit="page" from="458" to="462" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Regroupements issus de dépendancess syntaxiques en corpus : catégorisation et confontation à deux modélisations conceptuelles</title>
		<author>
			<persName><forename type="first">J</forename><surname>Bouaud</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Habert</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Nazarenko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Zweigenbaum</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Actes des Journées Ingénierie des Connaissances</title>
				<editor>
			<persName><forename type="first">E</forename><surname>Zacklad</surname></persName>
		</editor>
		<meeting>s des Journées Ingénierie des Connaissances<address><addrLine>Roscoff, France</addrLine></address></meeting>
		<imprint>
			<date type="published" when="1997-05">1997. May</date>
			<biblScope unit="page" from="207" to="223" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Word Association Norms, Mutual Information, and Lexicography</title>
		<author>
			<persName><forename type="first">K</forename><forename type="middle">W</forename><surname>Church</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Hanks</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of the 27th Annual Meeting of the Association for Computational Linguistics</title>
				<meeting>of the 27th Annual Meeting of the Association for Computational Linguistics</meeting>
		<imprint>
			<date type="published" when="1989">1989</date>
			<biblScope unit="page" from="76" to="83" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Similarity-Based Estimation of Word Co-occurrence Probabilities</title>
		<author>
			<persName><forename type="first">I</forename><surname>Dagan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Pereira</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Lee</forename><forename type="middle">L</forename></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 32nd Annual Meeting of the Association for Computational Linguistics, ACL&apos;94</title>
				<meeting>the 32nd Annual Meeting of the Association for Computational Linguistics, ACL&apos;94<address><addrLine>New Mexico State University</addrLine></address></meeting>
		<imprint>
			<date type="published" when="1994-06">1994. June</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Efficient Algorithms for Agglomerative Hierarchical Clustering Methods</title>
		<author>
			<persName><forename type="first">W</forename><surname>Day</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Edelsbrunner</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Classification</title>
		<imprint>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page" from="1" to="24" />
			<date type="published" when="1984">1984</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">A Corpus-based Conceptual Clustering Method for Verb Frames and Ontology Acquisition</title>
		<author>
			<persName><forename type="first">D</forename><surname>Faure</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Nédellec</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Workshop of the 1st Intl. Conf. on Language Resources and Evaluation</title>
				<editor>
			<persName><forename type="first">P</forename><surname>Velardi</surname></persName>
		</editor>
		<meeting><address><addrLine>Granada, Spain</addrLine></address></meeting>
		<imprint>
			<date type="published" when="1998-05">1998. May</date>
			<biblScope unit="page" from="1" to="8" />
		</imprint>
	</monogr>
	<note>Adapting lexical and corpus resources to sublanguages and applications</note>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Knowledge Acquisition via Incremental Conceptual Clustering</title>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">H</forename><surname>Fisher</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Machine Learning Journal</title>
		<imprint>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="page" from="139" to="172" />
			<date type="published" when="1987">1987</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Model of Incremental Concept Formation</title>
		<author>
			<persName><forename type="first">J</forename><surname>Gennari</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Langley</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Fisher</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Artificial Intelligence Journal</title>
		<imprint>
			<biblScope unit="volume">40</biblScope>
			<biblScope unit="page" from="11" to="61" />
			<date type="published" when="1989">1989</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Linking WordNet Verb Classes to Semantic Interpretation</title>
		<author>
			<persName><forename type="first">F</forename><surname>Gomez</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the COLING-ACL Workshop on the Usage of WordNet in NLP Systems</title>
				<meeting>the COLING-ACL Workshop on the Usage of WordNet in NLP Systems</meeting>
		<imprint>
			<date type="published" when="1998">1998</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Use of syntactic Context to Produce Term Association Lists for Text Retrieval</title>
		<author>
			<persName><forename type="first">G</forename><surname>Grefenstette</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 15th International SIGIR&apos;92</title>
				<meeting>the 15th International SIGIR&apos;92<address><addrLine>Denmark</addrLine></address></meeting>
		<imprint>
			<date type="published" when="1992">1992</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Evaluation Techniques for Automatic Semantic Extraction: Comparing Syntactic and Window Based Approaches</title>
		<author>
			<persName><forename type="first">G</forename><surname>Grefenstette</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Workshop on Acquisition of Lexical Knowledge form Text</title>
				<meeting><address><addrLine>Columbus, OH</addrLine></address></meeting>
		<imprint>
			<date type="published" when="1993-06">1993. June</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Generalizing Automaticaly Generated Selectional Patterns</title>
		<author>
			<persName><forename type="first">R</forename><surname>Grishman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Sterling</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 16th International Conference on Computational Linguistics (COLING&apos;94)</title>
				<meeting>the 16th International Conference on Computational Linguistics (COLING&apos;94)</meeting>
		<imprint>
			<date type="published" when="1994">1994</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Noun classification from predicate-argument structure</title>
		<author>
			<persName><forename type="first">D</forename><surname>Hindle</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of the 28th Annual Meeting of the Association for Computational Linguistics (ACL&apos;90)</title>
				<meeting>of the 28th Annual Meeting of the Association for Computational Linguistics (ACL&apos;90)<address><addrLine>Pittsburgh</addrLine></address></meeting>
		<imprint>
			<date type="published" when="1990">1990</date>
			<biblScope unit="page" from="268" to="275" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">The form of Information in Science</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Harris</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Gottfried</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Ryckman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Mattick</surname><genName>Jr</genName></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Daladier</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Harris</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Harris</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Analysis of Immunology Sublanguages</title>
		<title level="s">Boston Studies in the Philosophy of Science</title>
		<meeting><address><addrLine>Dordrecht, the Netherlands</addrLine></address></meeting>
		<imprint>
			<publisher>Kluwer Academic Publishers</publisher>
			<date type="published" when="1989">1989</date>
			<biblScope unit="volume">104</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Lexical Ambiguity and Information Retrieval</title>
		<author>
			<persName><forename type="first">R</forename><surname>Krovetz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><forename type="middle">B</forename><surname>Croft</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Lexical Acquisition: exploiting on-line resources to build a lexicon</title>
				<editor>
			<persName><surname>Zernik</surname></persName>
		</editor>
		<meeting><address><addrLine>Hillsdale, New Jersey</addrLine></address></meeting>
		<imprint>
			<publisher>Lawrence Erlbaum Associates</publisher>
			<date type="published" when="1991">1991</date>
			<biblScope unit="page" from="45" to="65" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Word clustering and disambiguation based on co-occurrence data</title>
		<author>
			<persName><forename type="first">H</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Abe</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of COLING -ACL&apos;98</title>
				<meeting>COLING -ACL&apos;98</meeting>
		<imprint>
			<date type="published" when="1998">1998</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">WordNet: an on-line lexical database</title>
		<author>
			<persName><forename type="first">G</forename><surname>Miller</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">International Journal of Lexicography</title>
		<imprint>
			<biblScope unit="volume">3</biblScope>
			<biblScope unit="issue">4</biblScope>
			<date type="published" when="1990">1990</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">Learning from Observation : Conceptual Clustering</title>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">S</forename><surname>Michalski</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Stepp</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Machine Learning I: an Artificial Intelligence Approach</title>
				<meeting><address><addrLine>Tioga</addrLine></address></meeting>
		<imprint>
			<date type="published" when="1983">1983</date>
			<biblScope unit="page" from="331" to="363" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">Distributional clustering of English words</title>
		<author>
			<persName><forename type="first">F</forename><surname>Pereira</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Tishby</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Lee</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 31st Annual Meeting of the Association for Computational Linguistics ACL&apos;93</title>
				<meeting>the 31st Annual Meeting of the Association for Computational Linguistics ACL&apos;93</meeting>
		<imprint>
			<date type="published" when="1993">1993</date>
			<biblScope unit="page" from="183" to="190" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">Concept based Query Expansion</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Qiu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><forename type="middle">P</forename><surname>Frei</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of 16th Annual International ACM SIGIR Conference</title>
				<meeting>16th Annual International ACM SIGIR Conference<address><addrLine>Pittsburgh</addrLine></address></meeting>
		<imprint>
			<publisher>ACM Press</publisher>
			<date type="published" when="1993">1993</date>
			<biblScope unit="page" from="160" to="169" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">Using Information Content to evaluate Semantic Similarity in a Taxonomy</title>
		<author>
			<persName><forename type="first">P</forename><surname>Resnik</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Cognitive Modelling</title>
				<imprint>
			<date type="published" when="1995">1995</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">Structural Ambiguity and Conceptual Relations</title>
		<author>
			<persName><forename type="first">P</forename><surname>Resnik</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">A</forename><surname>Hearst</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. Workshop on Very Large Corpora: Academic and Industrial Perspectives</title>
				<meeting>Workshop on Very Large Corpora: Academic and Industrial Perspectives<address><addrLine>Ohio State Univ</addrLine></address></meeting>
		<imprint>
			<date type="published" when="1993">1993</date>
			<biblScope unit="page" from="58" to="64" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">On Learning More Appropriate Selectional Restrictions</title>
		<author>
			<persName><forename type="first">F</forename><surname>Ribas</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of EACL&apos;95</title>
				<meeting>EACL&apos;95</meeting>
		<imprint>
			<date type="published" when="1995">1995</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<analytic>
		<title level="a" type="main">How Verb Subcategorization Frequencies Are Affected By Corpus Choice</title>
		<author>
			<persName><forename type="first">D</forename><surname>Roland</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Jurafsky</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Int&apos;l Conf. Computational Linguistics (COLING&apos;98)</title>
				<meeting>the Int&apos;l Conf. Computational Linguistics (COLING&apos;98)</meeting>
		<imprint>
			<date type="published" when="1998">1998</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b25">
	<analytic>
		<title level="a" type="main">Automatic Learning for Semantic Collocation</title>
		<author>
			<persName><forename type="first">S</forename><surname>Sekine</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">J</forename><surname>Caroll</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Ananiadou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Tsujii</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of the 3rd Conference on Applied Natural Language Processing</title>
				<meeting>of the 3rd Conference on Applied Natural Language essing</meeting>
		<imprint>
			<date type="published" when="1992">1992</date>
			<biblScope unit="page" from="104" to="109" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b26">
	<analytic>
		<title level="a" type="main">What makes an automatic kyewords classification effective?</title>
		<author>
			<persName><forename type="first">Sparck</forename><surname>Jones</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Barber</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">B</forename></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of the ASIS</title>
		<imprint>
			<biblScope unit="volume">18</biblScope>
			<biblScope unit="page" from="166" to="175" />
			<date type="published" when="1971">1971</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b27">
	<analytic>
		<title level="a" type="main">Efficient construction of comprehensible hierarchical clusterings</title>
		<author>
			<persName><forename type="first">L</forename><surname>Talavera</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Bejar</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2nd European Symposium on Principles of Data Mining and Knowledge Discovery, PKDD&apos;98</title>
		<title level="s">LNAI</title>
		<editor>
			<persName><forename type="first">France</forename><forename type="middle">J M</forename><surname>Nantes</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><surname>Zytkow</surname></persName>
		</editor>
		<editor>
			<persName><surname>Quafafou</surname></persName>
		</editor>
		<meeting>the 2nd European Symposium on Principles of Data Mining and Knowledge Discovery, PKDD&apos;98</meeting>
		<imprint>
			<publisher>Springer Verlag</publisher>
			<date type="published" when="1998">1998</date>
			<biblScope unit="volume">1510</biblScope>
			<biblScope unit="page" from="93" to="101" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b28">
	<analytic>
		<title level="a" type="main">A knowledge acquisition tool for multi-perspective concept formation</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">J F</forename><surname>Vasco</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Faicher</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Chouraqui</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">proceedings of 9th European Knowledge Acquisition Workshop, EKAW&apos;96</title>
				<meeting>9th European Knowledge Acquisition Workshop, EKAW&apos;96</meeting>
		<imprint>
			<publisher>Springer Verlag</publisher>
			<date type="published" when="1996">1996</date>
			<biblScope unit="page" from="227" to="244" />
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
