<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Designing clustering methods for ontology building: The Mo'K workbench</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Gilles Bisson</string-name>
          <email>Gilles.Bisson@imag.fr</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Claire Nédellec</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Dolores Cañamero</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>HELIX Project</institution>
          ,
          <addr-line>INRIA Rhône-Alpes, ZIRST, 655 Avenue de l'Europe</addr-line>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Inference and Learning Group, LRI</institution>
          ,
          <addr-line>Bât. 490, CNRS UMR 8623 &amp;</addr-line>
        </aff>
      </contrib-group>
      <abstract>
        <p>   This paper describes Mo'K, a configurable workbench that supports the development of conceptual clustering methods for ontology building. Mo'K is intended to assist ontology developers in the exploratory process of defining the most suitable learning methods for a given task. To do so, it provides facilities for evaluation, comparison, characterization and elaboration of conceptual clustering methods. Also, the model underlying Mo'K permits a finegrained definition of similarity measures and class construction operators, easing the tasks of method instantiation and configuration. This paper presents some experimental results that illustrate the suitability of the model to help characterize and assess the performance of different methods that learn semantic classes from parsed corpora.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. INTRODUCTION</title>
      <p>In this paper we propose a workbench that supports the
development of conceptual clustering methods for the (semi-)
automatic construction of ontologies of a conceptual hierarchy
type from parsed corpora. The elaboration of any clustering
method involves the definition of two main elements—a
distance metrics and a classification algorithm. In the context
of conceptual hierarchy formation, the Natural Language
Processing (NLP) community has investigated the notion of
distance to elaborate the semantic classes underlying
hierarchies. Classification algorithms have been broadly studied
within the Machine Learning and Data Analysis communities.</p>
      <p>
        Different tools have been developed for the automatic or
semi-automatic acquisition of semantic classes from “near”
terms. The notion of semantic proximity is based upon distance
among terms, defined as a function of the degree of similarity of
the contexts. Descriptions of term contexts (the learning
examples) and of the regularities to be sought vary in different
approaches. Contexts can be purely graphic—words
cooccurring within a window—as in the case of [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]; in some
cases, the window can cover the whole document (see e.g. [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ]).
Contexts can also be syntactic, as in the approaches that we
have taken into account to develop our model, e.g. [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ], [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ],
[
        <xref ref-type="bibr" rid="ref11">11</xref>
        ], [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ], [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], [
        <xref ref-type="bibr" rid="ref26">26</xref>
        ], [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. However, the selection of a suitable
distance for a given corpus and task is still an open problem
that has not received much attention so far [
        <xref ref-type="bibr" rid="ref25">25</xref>
        ]. In most cases,
the criteria proposed to support this choice rely on the
evaluation of the application task for which learning takes
place, as described for instance in [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. Evaluation criteria
proposed to assess learning results are purely quantitative, and
comparative analyses of these criteria are rare [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. A proper
characterization of the effects that different methods have on
the learning results would provide methodological guidelines
to help the designer select the most suitable method for a given
corpus and task, or to provide support to create a new one.
      </p>
      <p>
        This observation also applies to classification algorithms.
No meth odology or tool has been proposed to support the
elaboration of conceptual clustering algorithms that build
taskspecific ontologies. Work on conceptual clustering (e.g., [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ],
[
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ], [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], [
        <xref ref-type="bibr" rid="ref26">26</xref>
        ]) has not been extensively applied to the
problem of learning from corpora. One must however
acknowledge that the application of conceptual clustering
techniques to this domain is not straightforward, as existing
algorithms must be previously adapted. As in the case of
distances, the elaboration and selection of a suitable algorithm
for a given corpus and task requires the development of new
methodological guidelines and tools.
      </p>
      <p>As a first step toward this goal we propose Mo’K, a
configurable workbench to support the comparison, evaluation,
and elaboration of methods to learn conceptual hierarchies. The
conceptual clustering model underlying Mo’K permits a
finegrained definition of the components of distances and of class
construction operators, easing the tasks of method
instantiation and configuration. The model is extended with a
set of variables that permit to characterize features specific to
the elaboration of learning corpora, such as pruning, stop-lists,
etc. The workbench also includes evaluation criteria to assess
learning results obtained for different parameter configurations.
We finally present some experimental results that illustrate the
suitability of the model to help characterize different methods
and assess their performance. These results concern only class
formation, not classification algorithms.</p>
    </sec>
    <sec id="sec-2">
      <title>2. FRAMEWORK</title>
    </sec>
    <sec id="sec-3">
      <title>2.1 Learning semantic classes</title>
      <p>
        In the context of learning semantic classes, learning from
syntactic contexts exploits syntactic relations among words to
derive semantic relations, following Harris’ hypothesis [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ].
According to this hypothesis, the study of syntactic
regularities within a specialized corpus permits to identify
syntactic schemata made out of combinations of word classes
reflecting specific domain knowledge. The fact of using
specialized corpora eases the learning task, given that we have
to deal with a limited vocabulary with reduced polysemy, and
limited syntactic variability.
experiments, we depart from this practice to compare learned
classes, as we are interested in an extensional representation; we
therefore use classes formed by the union of attributes of near
objects (Figure 1.b).
      </p>
      <p>
        In syntactic approaches, learning results can be of different
types, depending on the method employed. They can be
distances that reflect the degree of similarity among terms [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ],
[
        <xref ref-type="bibr" rid="ref26">26</xref>
        ], [
        <xref ref-type="bibr" rid="ref22">22</xref>
        ], distance-based term classes elaborated with the help
of nearest-neighbor methods [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ], [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ], degrees of membership
in term classes [
        <xref ref-type="bibr" rid="ref24">24</xref>
        ], class hierarchies formed by conceptual
clustering [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ], or predicative schemata that use concepts to
constraint selection [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ], [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. The notion of distance is
fundamental in all cases, as it allows to calculate the degree of
proximity between two objects—terms in this case—as a
function of the degree of similarity between the syntactic
contexts in which they appear. Classes built by aggregation of
near terms can afterwards be used for different applications,
such as syntactic disambiguation [
        <xref ref-type="bibr" rid="ref23">23</xref>
        ], [
        <xref ref-type="bibr" rid="ref24">24</xref>
        ] or document
retrieval [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]. Distances are however calculated using the same
similarity notion in all cases, and our model relies on these
studies regardless of the application task.
      </p>
    </sec>
    <sec id="sec-4">
      <title>2.2 Conceptual Clustering</title>
      <p>
        In our case, ontologies are organized as multiple hierarchies
that form an acyclic graph where nodes are term categories
described by intention, and li nks represent inclusion, seen in
this case as a generality relation. Learning though hierarchical
classification of a set of objects can be performed in two main
ways: top-down, by incremental specialization of classes, and
bottom-up, by incremental generalization. We have adopted a
bottom-up approach due to its smaller algorithmic complexity,
and its understandability to the user in view of an interactive
validation task. In this article we focus on the elements needed
to build and evaluate the basic classes of this graph, i.e. criteria
for building the initial corpus, distances, and evaluation
criteria to asses results. We do not address the generic class
construction algorithm. With respect to this latter, let us just
mention that the application of hierarchical (conceptual [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ] or
numerical [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]) clustering algorithms to our problem is not
straightforward, given that we must build acyclic graphs with
few abstraction levels, rather than deep and strict hierarchies.
      </p>
    </sec>
    <sec id="sec-5">
      <title>3. THE MO’K MODEL AND WORKBENCH</title>
    </sec>
    <sec id="sec-6">
      <title>3.1  Representation of examples and results</title>
      <p>Following the standard practice, we use binary grammatical
relations as syntactic contexts. Examples are therefore
represented by triplets &lt;Head – Grammatical Relation –
Modifier head&gt;, where &lt;Modifier head&gt; is the object that must
be classified and &lt;Head – Grammatical Relation&gt; represents the
attribute. The number of occurrences of a triplet in a corpus
characterizes the attribute for an example. For instance, if we are
interested in verbal attachments, the following two sentences:
• This causes a decrease in […].</p>
      <p>• This high rate results from an increase in […].
allow to generate two triplets, &lt;cause Dobj decrease &gt; (29), and
&lt;Result from(Adj) increase &gt; (2), both presenting the structure
&lt;Verb – grammatical relation– Head noun&gt; (total number of
occurrences of these triads in the corpus). In the remaining of
the paper, we will designate by Action the tuple &lt;Verb –
grammatical relation&gt;. Actions &lt;Cause Dobj&gt; and &lt;Result
from(Adj)&gt; can be regarded as objects, and nouns &lt;Decrease&gt;
and &lt;Increase&gt; can be considered as attributes with values 29
and 2, respectively.</p>
      <p>In bottom-up clustering, couples of near objects or of
objects and classes are incrementally grouped in order to form
hierarchies or graphs of object classes. The standard in NLP is
to use object classes (Figure 1.a) for the application task. In our</p>
      <p>Let us take an example. If the objects &lt;Cause Dobj&gt; and &lt;Result
from(Adj)&gt; are selected to form a class, their attribute sets are
merged. Let us suppose that &lt;Cause Dobj&gt; is described by the
nouns {decrease, increase, modification, loss, etc.}, and
&lt;Result from(Adj)&gt; by {decrease, increase, composition,
evolution, etc.}. The noun class learned will include nouns
shared by both objects (in bold), and also the complementary
terms (in italics); therefore, four new triplets are induced. We
will then use the “attribute class” strategy, as th is way the “leaf”
classes that we will later evaluate will be larger than those
formed using the “object class” strategy. We will not further
develop here the differences between these two viewpoints,
intension and extension, since this topic is out of the scope of
the paper. Let us however insist on the fact that the selection of
one or the other has major effects on the learning results.</p>
    </sec>
    <sec id="sec-7">
      <title>3.2 Corpus parameters</title>
      <p>The parameters used to form a learning corpus in Mo’K include,
among others, selection of learning examples, level of pruning,
and “cleaning” of the corpus. Let us examine the first two.</p>
      <sec id="sec-7-1">
        <title>3.2.1 Selection of learning examples</title>
        <p>
          One of the goals of our model is to allow the user to compare
learning results as a function of the grammatical relations
selected as input. Objects and syntactic contexts used in
classification vary in different approaches—i.e. verbs or nouns
which are considered as similar on the grounds of their shared
verbal or nominal contexts, where nouns can be verb
complement heads (arguments [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ], or adjuncts [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ]), noun
complements [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ] or all of them [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ], [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ]. None of these
approaches proposes a comparative study of results based on
grammatical relations chosen in the initial corpus. Our model
easily allows to specify these relations. Experiments
concerning verbal relations reported below (Section 4.2)
illustrate this and show significant differences among results
depending on the nature of objects and attributes (whether they
are nouns or verbs), and on the type of corpus.
        </p>
      </sec>
      <sec id="sec-7-2">
        <title>3.2.2 Pruning</title>
        <p>A second parameter, taken into account by most existing
methods and included in our model, concerns corpus pruning
as a function of the number of occurrences of an element.
Pruning removes occurrences that are too infrequent and
therefore would cause noise, as well as those which are too
frequent and do not provide any information regarding the link
between an object and an attribute. The other side of the coin is
that infrequent but important cases can be removed. Our model
also allows to specify the minimum number of examples
characterizing an attribute and the minimal number of attributes
for an example and, for each of these constraints, the minimal
total number of occurrences of the triplets being considered.
The experiments reported in Section 4.3 show that the level of
pruning has a major impact on the results of learning, and that
the optimal level depends on the corpus.</p>
      </sec>
    </sec>
    <sec id="sec-8">
      <title>3.3 Distance Modeling</title>
      <p>Our goal is not to cover all the possible methods that can be
used to measure similarity between examples. On the contrary,
our approach focuses on methods with very precise features:
• They take syntactic analysis as input;
• They do not take into account external resources (e.g.</p>
      <p>
        ontologies such as WordNet [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ]);
• They are based on a comparison of the distribution profiles of
the attributes describing the couples of object to classify.
Different methods have been proposed in the NLP literature
with in this framework—among others [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ], [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ], [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] and
[
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. We have developed a generic model of these methods and
implemented it in Mo’K with the aim of elaborating a
comparison and evaluation methodology for them. In order to
come up with a generic implementation, we have identified the
steps shared by all these methods, as we will see below. Mo’K is
thus a workbench that implements a set of instantiable generic
methods using an object-oriented representation, as opposed to
the idea of a library of methods. This approach is made possible
by the fact that similarity measures can in general be regarded
as a comparison of the “distribution profiles” of couples of
examples. This way, two objects will be considered as
neighbors if the relative occurrence frequencies of each of their
attributes (i.e. of the syntactic contexts) are close. Learning
examples taken into account in our model can be represented by
means of a contingency table. Depending on the representation
hypothesis adopted, rows (examples) and columns (attributes)
of the table represent different things. For example, in the
experiments reported in Section 4.2 they first represent actions
and nouns, respectively, and nouns and actions later on. In any
case, a table cell contains the number of occurrences of an
attribute for a given example. This table is obviously very
sparse, as examples are generally described by a small number
of attributes (see Figure 2).
      </p>
      <p>In practice, computation of similarity can be decomposed
into two major st eps—weighting and similarity computation.
• The weighting phase changes every raw value of
cooccurrences appearing in the contingency table by a
coefficient, often normalized, which can be regarded as a
weight or measure of the significance of the fact of examples
and attributes co-occurring in the corpus. Its computation can
entail two steps—the initialization of the weights of
examples and attributes, usually according to their number of
occurrences, and the calculation of a normalized weight of the
relevance of each attribute for each example. Technically, this
weighting phase is implemented in Mo’K using the 5
functions described (in pseudo-code) in Table 1.
• The similarity computation phase builds a similarity matrix
between couples of examples. Similarity increases as a
function of the number of shared attributes, but the way in
which similarity between these distributions is calculated
varies in the different approaches. In Mo’K this phase is
implemented by a single function.</p>
      <p>We thus see that, by means of 6 functions and using a few lines
of code, it is possible to implement most similarity measures
that follow a schema based on comparison of pairs of
distribution profiles. Let us no te that we do not make more
specific hypotheses concerning the formal properties of
measures—they can be similarities or dissimilarities,
symmetrical or asymmetrical, and computed information can be
of any type. This approach thus favors the comparison of
existing methods, but also the elaboration of variants of these
methods and even the creation of new ones. Once integrated in
Mo'K, a method can access all the test and conceptual clustering
resources of the system.</p>
      <p>Name of the step
Initialization of the weight of each example E: W(E)
Initialization of the weight of each attribute A: W(A)
For each example E</p>
      <p>For each attribute A of the example</p>
      <p>Calculate W(A) in the context of E
Update global W(E)
For each attribute A of the example</p>
      <p>Normalization of the W(A) by W(E)</p>
      <p>Method
Init_Weight_Example
Init_Weight_Attribute
Eval_Weight_Example
Eval_Weight_Attribute
Init_Similarity</p>
    </sec>
    <sec id="sec-9">
      <title>3.4 Distance evaluation</title>
      <p>Even though our goal is the construction of hierarchies, it is
interesting to evaluate the relevance of a distance metrics with
respect to more simple tasks and to analyze its behavior as a
function of the application domain and of the parameters of
elaboration of the learning corpus. Mo'K offers different means
of evaluation based on the first N couples of examples built by
binary aggregation, i.e. the first N couples of examples with
highest scores in the similarity matrix.</p>
      <sec id="sec-9-1">
        <title>3.4.1 Measure of recall</title>
        <p>As already mentioned in Section 3, the elaboration of a class
gives rise to the induction of new triplets not observed in the
initial corpus. Therefore, the evaluation process follows the
classical schema of dividing the corpus in two
partitions—learning and test. The former is used to build the
similarity matrix according to the measure to be evaluated. The
latter allows to measure the coverage rates of classes, i.e. their
ability to recognize the triplets in the test set. We have adopted
this evaluation task for two reasons. First, it corresponds to the
elementary step in every process of bottom-up hierarchical
clustering. Second, from a NLP perspective, it conforms to a
disambiguation task.</p>
      </sec>
      <sec id="sec-9-2">
        <title>3.4.2 Measure of precision</title>
        <p>
          Despite its interest, the coverage measure only allows to
evaluate the recall rate associated with the set N of selected
classes. However, precision—a measure of the ability to avoid
erroneous recognition of negative examples—is an equally
important property of the metrics. In the end, a similarity
measure that tends to over-generalize and describe object
couples using a large number of attributes would reach high
coverage rates, but produce classes that lack in meaning and
precision. It is difficult to automatically solve the problem of
evaluating unsupervised learning in the absence of negative
examples. Given that we do not deal with annotated corpora,
and we do not have negative examples, we face this problem in
Mo’K by means of automatically generated (artificial) sample
corpora. Following [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ], we assume that examples generated this
way will be negatives for the most part. Artifical examples are
formed by randomly choosing an object and an attribute from
the initial corpus, taking care that none of these examples
appears in the learning set. We measure coverage rates on
artificial examples using learned classes. Since examples are
randomly generated, some positive examples are generated as
well (about 0.5% in the studied corpora). Although this rate of
artificial positive examples might seem very low, it
unexpectedly constitutes an important part of the artificial
examples covered by learned classes—they cover between 0.5%
and 2.5% artificial examples in our experiments. Hence, real
precision can only be evaluated after negative examples in the
artificially generated set have been computed by hand.
        </p>
        <p>As we will see in the experiments reported in next section, it
is interesting to measure other criteria in order to assess the
relevance of a similarity measure—for example, the induction
rate measuring the ratio between the number of induced triplets
and the total number of triplets learned.</p>
      </sec>
    </sec>
    <sec id="sec-10">
      <title>4. EXPERIMENTS AND RESULTS</title>
      <p>The experiments reported here aim to illustrate Mo'K’s
parameterization possibilities and the impact that different
parameter settings have on the learning results. These
experiments make thus a case for the use of generic platforms to
perform a systematic exploratory analysis in order to obtain
sensible results in a given domain (corpus).</p>
    </sec>
    <sec id="sec-11">
      <title>4.1 Training corpora</title>
      <p>We have conducted experiments on two different French
corpora—one contains cooking recipes gathered over the world
wide web; the other, Agrovoc, contains scientific abstracts in
the agricultural domain, and has been assembled by the INIST
(Institut de l’Information Scientifique et Technique) of the
CNRS. We have chosen these two corpora as they differ in
generality and amount of technical terms, but are still close
enough to allow for meaningful comparison of results. Both
corpora have been analyzed using the same shallow parser. Only
verbal relations of the form &lt;Verb – Grammatical Relation –
Noun&gt; have been considered in our experiments. The output of
syntactic parsing is highly noisy (between 30% and 50%
mistakes) due to several factors such as grammatical and
spelling mistakes, typos, and accentuation errors in the case of
the cooking corpus. In Agrovoc, noise is mostly produced by
the high number of technical terms mistaken by verbs, and of
embedded noun complements which are erroneously attached to
verbs. In Agrovoc, only 300 verbs are found versus 18,828
nouns. This is due to the fact that only part of the corpus has
been considered—those triplets with a verb that appears in a
list of verbs giving rise to nominalizations in the corpus. In the
cooking corpus we find 1,181 verbs for 3,300 nouns, i.e. the
ratio between nouns and verbs is, in average, divided by a factor
of 20 with respect to Agrovoc. This reflects a higher
specialization in this corpus. Finally, Agrovoc is three times as
big as the cooking corpus (117,156 triplets for 168,287
occurrences in total).</p>
    </sec>
    <sec id="sec-12">
      <title>4.2 Selection of example representation</title>
      <p>The first experiment illustrates the importance of the choice of
the object to be clustered and of the attribute which
characterizes it.We have compared the classes learned by
considering actions as objects and head nouns as attributes
(denoted as action-based representation), with those learned
using nouns as objects and actions as attributes (denoted as
noun-based representation). The comparison (see Table 2) has
been performed on both corpora, the other parameters remaining
unchanged. We have applied Asium’s distance. Pruning criteria
are light—no minimum number of triplet occurrences, the
minimum number of examples characterizing a given attribute
has been set to 2, and the minimum number of attributes for an
example to 3. We will further comment this setting in Section
4.3. For each corpus, the first 25 learned classes are evaluated.
Coverage is measured on the test set, which comprises 20% of
the whole corpus, and on the artificial set, which contains
50,000 triplets randomly generated (see Section 3.3). Each test
has been repeated four times.</p>
      <p>The first experiment has been conducted on Agrovoc. The
classes learned in the action-based representation are twice as
large as those learned in the noun-based representation.
However, induction rates (number of induced triplets divided
by the number of triplets learned by rote) are very similar (40%
compared to 38%) and so are precision rates (45% in both
cases). Precision rate represents the rate of negative examples
among learned examples which cover artificial examples. The
recall measured by the coverage rate of induced triplets on the
test set is slightly better for the noun-based representation
(5.3% compared to 4.7%) but remains quite low. This can be
explained by the level of generality of what is learned: the best
couples of learned nouns involved very general terms such as
[technique-method], [influence-effect], in contrast with the
numerous technical words of the corpus. Most of the actions
characterizing these nouns concern general verbs such as [to
present], [to observe], and [to report]. This is confirmed by
looking at the best pairs of actions such as &lt;to study Dobj&gt;
&lt;to analyze Dobj&gt;. This explains why learned classes are of
rather poor quality in both representations.</p>
      <p>Corpus
Agrovoc
Cooking</p>
      <p>Learning
object
Action
Nom
Action
Nom
% Induced tripl./ Recall (test Precision
learned tripl. set)
40 % 4.7 % 45 %
38 % 5.3 % 45 %
34 % 12 % 32 %
38 % 9.1 % 52 %</p>
      <p>The experiments on the cooking corpus have built classes of
similar size in both representations. Induction rate is slightly
higher for the noun-based representation—38% compared to
34%. However, recall on the test set is better in the action-based
representation, (12% and 9.1%, respectively). Induced triplets
are thus more useful in the case of action-based representation,
even though they are less numerous. Moreover, the precision
measured by the rate of negative artificial examples covered by
learned examples is much better for the action-based
representation (32% compared to 52%). Precision and recall
rates are thus better in this representation, although the rate of
induced triplets is smaller than in the noun-based
representation. In any case, all rates are much better than the
ones computed for the Agrovoc experiments. A closer
examination of the best pairs of actions and nouns confirms the
idea that overgeneralization is less of a problem here than in
Agrovoc. Noun pairs are more precise (e.g., [fridge-freezer],
[olive oil–oil]) and described by more technical actions. In the
same way, the best pairs of actions, such as [absorb Dobj ,
evaporate Dobj ], are characterized by nouns (in this case
[vinegar, water, wine, excess, etc.]) which are significantly more
specific than in Agrovoc. The smaller variability of the cooking
corpus explains these observations, showing that the larger size
of Agrovoc does not improve the meaningfulness of the
regularities observed.</p>
      <p>This experiment thus shows that the choice of a
representation can have a major impact on the learning results.
It is therefore advisable to select a suitable representation
before addressing a new domain.</p>
    </sec>
    <sec id="sec-13">
      <title>4.3 Pruning parameters</title>
      <p>To illustrate the importance of pruning, two pruning settings
have been applied to both corpora and compared. In both cases
we have used Asium’s distance, objects are actions and
attributes are nouns. The first setting is the one described in the
previous section. In the second one, we have set the minimum
number of occurrences for an example (triplet) to 2, in order to
remove triplets occurring only once in the corpus, as they may
represent noise. We have also augmented the values of the
minimal number of examples that must characterize an attribute
(from 2 to 3), and of the minimal number of attributes per object
(from 3 to 5, i.e. in this version the objects being compared
appear in at least 5 different syntactic contexts). As it can be
inferred from the histogram in Fig. 2, this setting excludes 80%
of the corpus, versus 70% in the first setting. We can thus hope
for a more reliable classification, to the risk of removing so
many examples that coverage (recall) is drastically affected. The
experiments show that, for the cooking corpus, induction rate
(32%) and coverage (11.2%) on the whole test set are nearly
unaffected. On the contrary, the rate of artificial triplets covered
scales by a factor of 3 with respect to the previous rate; we think
that this significant increase indicates that this pruning rate
increases the rate of erroneously induced examples. In Agrovoc,
induction rate decreases by a factor of three and coverage rate
by a factor of two, whereas the rate of artificial triplets covered
scales by a factor superior to 2. Recall is therefore much
strongly affected by pruning than in the cooking corpus. In
both cases, this new pruning setting gives rise to a decrease in
performance.</p>
      <p>However, this conclusion drawn from the evaluation of basic
classes must be tempered in the case of hierarchy formation. In
this case, the more constraining version of pruning allows to
eliminate many non-significant classes that result from the
presence of closely similar actions described by a small number
of attributes. It seems clear that, in a process of hierarchical
clustering, this type of class would cause problems, as it would
alter some groupings. Therefore, the type of pruning to be
applied partly depends on the task to be performed.</p>
    </sec>
    <sec id="sec-14">
      <title>4.4 Comparison of methods</title>
      <p>
        This last experiment illustrates some aspects of the use of Mo'K
to compare results obtained with different distances. Among the
methods that we have tested with Mo’K (such as those proposed
by Dagan et al ., Hindle, Grefenstette, and Grishman et al . among
the best known in the literature, as well as other distances
proposed by the authors, e.g. Asium, and Greedy) we have
chosen to compare here the distances used in Asium [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], the one
proposed by Dagan [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], and Greedy3. The reason for th is choice
relies on the fact that these three methods have shown good
t3hGerχe2e,dwyhhiacsh afasviomrpslethbeeshealveicotiro:nitoifs bpaaisresd oofn eaxammepalseusred,eisncsrpibireedd fbryoma
large number of attributes (the method is named after this feature).
performance when compared with others, while presenting
rather different behaviors. They are thus a representative sample
of the subset of methods that we have modeled as characterized
in Section 3.3. These methods have been applied to the corpus
of cooking recipes. Learning parameters are the same as those
described in Section 4.2, objects are actions and attributes are
names. In addition, we have tested the influence on learning of
the number of disjointed classes on which recall performance is
evaluated. To do this, we have varied this number (in abscissa
of the diagrams below) between 10 and 200.
      </p>
      <p>These diagrams show, respectively, the recall rate of each
method on the test set (Figure 4a), and the efficiency of classes
rate (Figure 4b). The latter is assessed by the ratio between the
number of triplets learned (by rote and induced), and the
number of triplets effectively used in the recall test. As we can
see in the first diagram, the coverage rate of the three methods
grows as expected according to the number of classes
considered, Dagan’s method yielding the best results. On the
contrary, if we pay attention to the efficiency of classes, Asium
takes better advantage of the triplets learned. Looking at both
diagrams, we can conclude that these methods have different
behaviors. Dagan’s gives rise to more general classes (more
triplets are learned, but the number of useless ones is higher).
Asium constructs more specific classes (fewer triplets learned,
but more of them are useful). We can take a closer look at the
behavior of these methods and study the quality of induction
in terms of the rate of induced triplets which are effectively
used in the recall test (Figure 5).</p>
      <p>While the previous conclusion is confirmed for Asium and
Dagan’s methods, we can also note that Greedy and Dagan’s
methods have the same behavior along this criterion. In fact, it
seems that Dagan's method is able to induce more useful
triplets than Greedy, whereas this latter tends to learn by rote a
more representative sample subset of the corpus.</p>
      <p>The tests performed on the artificial examples confirm these
results. Therefore, it seems that the classes learned by Dagan's
method and, to a lesser extent, by Greedy are less robust and
present lower precision rates for of the learning parameters
chosen. We must emphasize that these conclusions only apply
to the recipe corpus. For Agrovoc, results are considerably
different. These experiments therefore show the importance of
going through an exploratory process in order to come up with
the most suitable methods and representation.</p>
    </sec>
    <sec id="sec-15">
      <title>5. CONCLUSIONS AND PROSPECTS</title>
      <p>
        Mo’K is a configurable workbench that supports the
development of conceptual clustering methods for specific
ontology building. The learning model proposed here takes
parsed corpora as input. No additional (terminological or
semantic) knowledge is used for labeling the input, guiding
learning, or validating the learning results. Preliminary
experiments showed that the quality of learning decreases with
the generality of the corpus. This makes somehow unrealistic
the use of general ontologies for guiding such learning, as they
seem too incomplete and polysemic to allow for efficient
learning in specific domains. For example, [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ] points out that
40% of the words in canonical form in the titles and abstracts of
the Communications of the ACM are not included in the LDOCE
(Longman Dictionary of Contemporary English). This problem
posed in the case of learning specific ontologies obviously
does no t apply in the case of guiding learning of general
semantic classes, as shown in the abundant literature on the
topic (see e.g. [
        <xref ref-type="bibr" rid="ref23">23</xref>
        ], [
        <xref ref-type="bibr" rid="ref22">22</xref>
        ], [
        <xref ref-type="bibr" rid="ref24">24</xref>
        ], [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ]). It would however be highly
valuable to take advantage of existing ontologies to improve
the quality of learning. We consider that this can be achieved in
two main ways. First, learning could be improved by the use of
specific terminologies, dictionaries and nomenclature, such as
SNOMED International in the medical domain [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. Second, some
methodological guidelines would be needed to integrate
specialized learned ontologies into more general ontologies
such as WordNet [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ].
      </p>
      <p>Although we have focused on a disambiguation-based task,
other validation tasks could be integrated in Mo’K, such as
query extension and information extraction. Learning
specialized ontologies of high quality for these tasks will allow
the development of applications in technical and rapidly
evolving domains, in which manual acquisition is too costly.
In this sense, we have started exploring information extraction
from molecular biology abstracts.</p>
    </sec>
    <sec id="sec-16">
      <title>ACKNOWLEDGEMENTS</title>
      <p>We are grateful to INIST-CNRS for providing the Agrovoc
corpus. This research is partly funded by the French Ministry of
Industry under RNRT project Astuxe.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>Basili</surname>
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pazienza</surname>
            <given-names>M. T.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Velardi</surname>
            <given-names>P.</given-names>
          </string-name>
          <year>1996</year>
          .
          <article-title>An empirical symbolic approach to natural language processing</article-title>
          .
          <source>Artificial Intelligence Journal 85</source>
          , pp.
          <fpage>59</fpage>
          -
          <lpage>99</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <surname>Bisson</surname>
            <given-names>G.</given-names>
          </string-name>
          <year>1992</year>
          .
          <article-title>Conceptual Clustering in a First Order Logic Representation</article-title>
          .
          <source>In Proceedings of 10th European Conference on Artificial Intelligence (ECAI'92)</source>
          , pp.
          <fpage>458</fpage>
          -
          <lpage>462</lpage>
          , Vienna .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <surname>Bouaud</surname>
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Habert</surname>
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nazarenko</surname>
            <given-names>A.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Zweigenbaum</surname>
            <given-names>P.</given-names>
          </string-name>
          <year>1997</year>
          .
          <article-title>Regroupements issus de dépendancess syntaxiques en corpus : catégorisation et confontation à deux modélisations conceptuelles</article-title>
          .
          <source>In Actes des Journées Ingénierie des Connaissances</source>
          , Zacklad E. (Ed.), pp.
          <fpage>207</fpage>
          -
          <lpage>223</lpage>
          , Roscoff, France, May.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <surname>Church</surname>
            <given-names>K. W.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Hanks</surname>
            <given-names>P.</given-names>
          </string-name>
          <year>1989</year>
          . Word Association Norms, Mutual Information, and Lexicography,
          <source>in Proc. of the 27th Annual Meeting of the Association for Computational Linguistics</source>
          , pp.
          <fpage>76</fpage>
          -
          <lpage>83</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <surname>Dagan</surname>
            <given-names>I.</given-names>
          </string-name>
          , Pereira F.,
          <string-name>
            <surname>and Lee L.</surname>
          </string-name>
          <year>1994</year>
          .
          <article-title>Similarity-Based Estimation of Word Co-occurrence Probabilities</article-title>
          .
          <source>In Proceedings of the 32nd Annual Meeting of the Association for Computational Linguistics</source>
          , ACL'
          <fpage>94</fpage>
          , New Mexico State University, June.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <surname>Day</surname>
            <given-names>W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Edelsbrunner</surname>
            <given-names>H.</given-names>
          </string-name>
          <year>1984</year>
          .
          <article-title>Efficient Algorithms for Agglomerative Hierarchical Clustering Methods</article-title>
          ,
          <source>Journal of Classification</source>
          . Volume
          <volume>1</volume>
          . pp.
          <fpage>1</fpage>
          -
          <lpage>24</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>Faure</surname>
            <given-names>D.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Nédellec</surname>
            <given-names>C.</given-names>
          </string-name>
          <year>1998</year>
          .
          <article-title>A Corpus-based Conceptual Clustering Method for Verb Frames and Ontology Acquisition</article-title>
          . In P. Velardi (Ed.),
          <article-title>Adapting lexical and corpus resources to sublanguages and applications</article-title>
          ,
          <source>Workshop of the 1st Intl. Conf. on Language Resources and Evaluation</source>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>8</lpage>
          , Granada, Spain, May.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <surname>Fisher</surname>
            <given-names>D.H.</given-names>
          </string-name>
          <year>1987</year>
          .
          <article-title>Knowledge Acquisition via Incremental Conceptual Clustering</article-title>
          ,
          <source>Machine Learning Journal 2</source>
          , pp. 
          <fpage>139</fpage>
          -
          <lpage>172</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <surname>Gennari</surname>
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Langley</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fisher</surname>
            <given-names>D.</given-names>
          </string-name>
          <year>1989</year>
          .
          <article-title>Model of Incremental Concept Formation</article-title>
          ,
          <source>Artificial Intelligence Journal</source>
          , Volume
          <volume>40</volume>
          ,
          <fpage>11</fpage>
          -
          <lpage>61</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>Gomez</surname>
            <given-names>F.</given-names>
          </string-name>
          <year>1998</year>
          .
          <article-title>Linking WordNet Verb Classes to Semantic Interpretation</article-title>
          .
          <source>In Proceedings of the COLING-ACL Workshop on the Usage of WordNet in NLP Systems.</source>
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>Grefenstette</surname>
            <given-names>G.</given-names>
          </string-name>
          <year>1992</year>
          .
          <article-title>Use of syntactic Context to Produce Term Association Lists for Text Retrieval</article-title>
          .
          <source>In Proceedings of the 15th International SIGIR'92</source>
          ,
          <string-name>
            <surname>Denmark</surname>
          </string-name>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <surname>Grefenstette</surname>
            <given-names>G.</given-names>
          </string-name>
          <year>1993</year>
          .
          <article-title>Evaluation Techniques for Automatic Semantic Extraction: Comparing Syntactic and Window Based Approaches</article-title>
          .
          <source>In Workshop on Acquisition of Lexical Knowledge form Text</source>
          , Columbus,
          <string-name>
            <surname>OH</surname>
          </string-name>
          , June.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <surname>Grishman</surname>
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sterling</surname>
            <given-names>J.</given-names>
          </string-name>
          <year>1994</year>
          .
          <article-title>Generalizing Automaticaly Generated Selectional Patterns</article-title>
          .
          <source>In Proceedings of the 16th International Conference on Computational Linguistics (COLING'94).</source>
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <surname>Hindle</surname>
            <given-names>D.</given-names>
          </string-name>
          <year>1990</year>
          .
          <article-title>Noun classification from predicate-argument structure</article-title>
          .
          <source>In Proc. of the 28th Annual Meeting of the Association for Computational Linguistics (ACL'90)</source>
          , pp.
          <fpage>268</fpage>
          -
          <lpage>275</lpage>
          , Pittsburgh.
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <surname>Harris</surname>
            <given-names>Z.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gottfried</surname>
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ryckman</surname>
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mattick</surname>
            Jr
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Daladier</surname>
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Harris</surname>
            <given-names>T.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Harris</surname>
            <given-names>S.</given-names>
          </string-name>
          <year>1989</year>
          .
          <article-title>The form of Information in Science</article-title>
          .
          <source>In Analysis of Immunology Sublanguages</source>
          , vol.
          <volume>104</volume>
          of Boston Studies in the Philosophy of Science. Dordrecht, the Netherlands, Kluwer Academic Publishers.
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <surname>Krovetz</surname>
            <given-names>R</given-names>
          </string-name>
          and Croft W.B.,
          <string-name>
            <surname>W.</surname>
          </string-name>
          <year>1991</year>
          .
          <article-title>Lexical Ambiguity and Information Retrieval. In Lexical Acquisition: exploiting on-line resources to build a lexicon ¿ Zernik (Ed</article-title>
          .), pp.
          <fpage>45</fpage>
          -
          <lpage>65</lpage>
          , Hillsdale, New Jersey, Lawrence Erlbaum Associates .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <surname>Li</surname>
            <given-names>H.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Abe</surname>
            <given-names>N.</given-names>
          </string-name>
          <year>1998</year>
          .
          <article-title>Word clustering and disambiguation based on co-occurrence data</article-title>
          .
          <source>In Proceedings of COLING - ACL'98.</source>
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <surname>Miller</surname>
            <given-names>G.</given-names>
          </string-name>
          <year>1990</year>
          .
          <article-title>WordNet: an on-line lexical database</article-title>
          ,
          <source>International Journal of Lexicography</source>
          ,
          <volume>3</volume>
          (
          <issue>4</issue>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <surname>Michalski</surname>
            <given-names>R.S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Stepp</surname>
            <given-names>E.</given-names>
          </string-name>
          <year>1983</year>
          .
          <article-title>Learning from Observation : Conceptual Clustering</article-title>
          .
          <source>In Machine Learning I: an Artificial Intelligence Approach</source>
          , Tioga, pp.
          <fpage>331</fpage>
          -
          <lpage>363</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <surname>Pereira</surname>
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tishby</surname>
            <given-names>N.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Lee</surname>
            <given-names>L.</given-names>
          </string-name>
          <year>1993</year>
          .
          <article-title>Distributional clustering of English words</article-title>
          .
          <source>In Proceedings of the 31st Annual Meeting of the Association for Computational Linguistics ACL'93</source>
          , p.
          <fpage>183</fpage>
          -
          <lpage>190</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <surname>Qiu</surname>
            <given-names>Y.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Frei</surname>
            <given-names>H. P.</given-names>
          </string-name>
          <year>1993</year>
          .
          <article-title>Concept based Query Expansion</article-title>
          .
          <source>In Proceedings of 16th Annual International ACM SIGIR Conference</source>
          , pp.
          <fpage>160</fpage>
          -
          <lpage>169</lpage>
          , Pittsburgh, ACM Press.
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <surname>Resnik</surname>
            <given-names>P.</given-names>
          </string-name>
          <year>1995</year>
          .
          <article-title>Using Information Content to evaluate Semantic Similarity in a Taxonomy, Cognitive Modelling</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <surname>Resnik</surname>
            <given-names>P.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Hearst</surname>
            <given-names>M. A.</given-names>
          </string-name>
          <year>1993</year>
          .
          <article-title>Structural Ambiguity and Conceptual Relations</article-title>
          .
          <source>In Proc. Workshop on Very Large Corpora: Academic and Industrial Perspectives</source>
          , pp.
          <fpage>58</fpage>
          -
          <lpage>64</lpage>
          , Ohio State Univ.
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <surname>Ribas</surname>
            <given-names>F.</given-names>
          </string-name>
          <year>1995</year>
          .
          <article-title>On Learning More Appropriate Selectional Restrictions</article-title>
          .
          <source>In Proceedings of EACL'95.</source>
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <surname>Roland</surname>
            <given-names>D.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Jurafsky</surname>
            <given-names>D.</given-names>
          </string-name>
          <year>1998</year>
          .
          <article-title>How Verb Subcategorization Frequencies Are Affected By Corpus Choice</article-title>
          .
          <source>In Proceedings of the Int'l Conf. Computational Linguistics (COLING'98).</source>
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [26]
          <string-name>
            <surname>Sekine</surname>
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Caroll</surname>
            <given-names>J. J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ananiadou</surname>
            <given-names>S.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Tsujii</surname>
            <given-names>J.</given-names>
          </string-name>
          <year>1992</year>
          .
          <article-title>Automatic Learning for Semantic Collocation</article-title>
          .
          <source>In Proc. of the 3rd Conference on Applied Natural Language Processing</source>
          , pp.
          <fpage>104</fpage>
          -
          <lpage>109</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [27]
          <string-name>
            <surname>Sparck Jones</surname>
            <given-names>K.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Barber</surname>
            <given-names>E. B.</given-names>
          </string-name>
          <year>1971</year>
          .
          <article-title>What makes an automatic kyewords classification effective?</article-title>
          ,
          <source>Journal of the ASIS</source>
          ,
          <volume>18</volume>
          :
          <fpage>166</fpage>
          -
          <lpage>175</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          [28]
          <string-name>
            <surname>Talavera</surname>
            <given-names>L.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Bejar</surname>
            <given-names>J.</given-names>
          </string-name>
          <year>1998</year>
          .
          <article-title>Efficient construction of comprehensible hierarchical clusterings</article-title>
          .
          <source>In Proceedings of the 2nd European Symposium on Principles of Data Mining and Knowledge Discovery</source>
          ,
          <source>PKDD'98</source>
          . pp.
          <fpage>93</fpage>
          -
          <lpage>101</lpage>
          . Nantes,
          <string-name>
            <given-names>France. J. M.</given-names>
            <surname>Zytkow</surname>
          </string-name>
          and M. Quafafou (eds.) LNAI vol.
          <volume>1510</volume>
          , Springer Verlag.
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          [29]
          <string-name>
            <surname>Vasco</surname>
            <given-names>J. J. F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Faicher</surname>
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chouraqui</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          <year>1996</year>
          .
          <article-title>A knowledge acquisition tool for multi-perspective concept formation</article-title>
          .
          <source>In proceedings of 9th European Knowledge Acquisition Workshop</source>
          , EKAW'
          <volume>96</volume>
          , pp.
          <fpage>227</fpage>
          -
          <lpage>244</lpage>
          . Springer Verlag.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>