<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>PrOntoLearn: Unsupervised Lexico-Semantic Ontology Generation using Probabilistic Methods</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Saminda Abeyruwan</string-name>
          <email>saminda@cs.miami.edu</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ubbo Visser</string-name>
          <email>visser@cs.miami.edu</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Vance Lemmon</string-name>
          <email>vlemmon@miami.edu</email>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Stephan Schurer</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Computer Science, University of Miami</institution>
          ,
          <addr-line>Florida</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Department of Molecular and Cellular Pharmacology, University of Miami Miller School of Medicine</institution>
          ,
          <addr-line>Florida</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>The Miami Project to Cure Paralysis, University of Miami Miller School of Medicine</institution>
          ,
          <addr-line>Florida</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Formalizing an ontology for a domain manually is well-known as a tedious and cumbersome process. It is constrained by the knowledge acquisition bottleneck. Therefore, researchers developed algorithms and systems that can help to automatize the process. Among them are systems that include text corpora for the acquisition. Our idea is also based on vast amount of text corpora. Here, we provide a novel unsupervised bottom-up ontology generation method. It is based on lexico-semantic structures and Bayesian reasoning to expedite the ontology generation process. We provide a quantitative and two qualitative results illustrating our approach using a high throughput screening assay corpus and two custom text corpora. This process could also provide evidence for domain experts to build ontologies based on top-down approaches.</p>
      </abstract>
      <kwd-group>
        <kwd>Ontology Modeling</kwd>
        <kwd>Ontology Learning</kwd>
        <kwd>Probabilistic Methods</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        An ontology is a formal, explicit speci cation of a shared conceptualization [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ],
[
        <xref ref-type="bibr" rid="ref22">22</xref>
        ]. Formalizing an ontology for a given domain with the supervision of domain
experts is a tedious and cumbersome process. The identi cation of the
structures and the characteristics of the domain knowledge through an ontology is a
demanding task. This problem is known as the knowledge acquisition bottleneck
(KAB) and a suitable solution presently does not exist.
      </p>
      <p>There exists a large number of text corpora available from di erent domains
(e.g., the BioAssay high throughput screening assays4) that need to be classi ed
into ontologies to faciliate the discovery of new knowledge. A domain of discourse</p>
    </sec>
    <sec id="sec-2">
      <title>4 http://bioassayontology.org/</title>
      <p>(i.e., sequential number of sentences) shows characteristics such as 1) redundancy
2) structured and unstructured text 3) noisy and uncertain data that provide a
degree of belief 4) lexical disambiguity, and 5) semantic heterogeneity problems.
We discuss in depth the importance of these characteristics in section 3. Our
goal in this research is to provide a novel method to construct an ontology from
the evidence collected from the corpus. In order to achieve our goal, we use the
lexico-semantic features of the lexicon and probabilistic reasoning to handle the
uncertainty of features. Since our method is applied to build an ontology for
a corpus without domain experts, this method can be seen as an unsupervised
learning technique. Since the method starts from the evidence present in the
corpus, it is can be seen as a reverse engineering technique. We use WordNet5 to
handle lexico-semantic structures, and the Bayesian reasoning to handle degree
of belief of an uncertain event. We implement a Java based application to serialize
the learned conceptualization to OWL DL6 format.</p>
      <p>The rest of the paper is organized as follows: section 2 provides a broad
investigation of the related work. Section 3 provides details of our research approach.
Section 4 provides a detail description of the experiments based on three
different text corpora and the discussion. Finally, section 5 provides the summary
and the future work.
2</p>
      <sec id="sec-2-1">
        <title>Related Work</title>
        <p>The problem of learning a conceptualization from a corpus has been studied in
many disciplines such as machine learning, text mining, information retrieval,
natural language processing, and Semantic Web. Table 1 shows the pros and cons
of di erent techniques to solve the problem of ontology learning. Each method
covers some portion of the problem and each method learns the conceptualization
from terms, and present it as taxonomies and axioms to an ontology. On the other
hand, most of the methods use a top-down approach, i.e., an initial classi cation
of an ontology is given. The uncertainty inherited from the domain is usually
dealt with by a domain expert, and the conceptualization is normally de ned
using prede ned rules or templates. These methods show the characteristics of
a semi-supervised and a semi-automated learning paradigm.
3</p>
      </sec>
      <sec id="sec-2-2">
        <title>Approach</title>
        <p>
          Our research focuses on an unsupervised method to quantify the degree of belief
that a grouping of words in the corpus will provide a substantial
conceptualization of the domain of interest. The degree of belief in world states in uences
the uncertainty of the conceptualization. The uncertainty arises from partial
observability, non-determinism, laziness and theoretical and practical ignorance
[
          <xref ref-type="bibr" rid="ref19">19</xref>
          ]. The partial observability arises from the size of the corpus. Even though
        </p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>5 http://wordnet.princeton.edu/</title>
    </sec>
    <sec id="sec-4">
      <title>6 http://www.w3.org/TR/owl-guide/</title>
      <p>
        PR [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ], [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ], [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] and [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ] reasoning available available prob. theory
NELL [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] 24 7 learning xed dynamic ML techniques
DART [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] world knowledge semi-automated
RTE [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], and [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] entailment ATP
NLU [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ] commonsense rules semi-supervised
      </p>
      <sec id="sec-4-1">
        <title>Text2Onto [6] ontology learning p p semi-supervised</title>
      </sec>
      <sec id="sec-4-2">
        <title>LexO [24] complex classes p semi-supervised</title>
      </sec>
      <sec id="sec-4-3">
        <title>FCA [5] taxonomy p FCA</title>
        <p>
          OP [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ], and [
          <xref ref-type="bibr" rid="ref23">23</xref>
          ] ontology population available available semi-/supervised
a corpus many be large, it might not contain all the necessary evidence of an
event of interest. A corpus contains ambiguous statements about an event that
leads to a non-determinism of the state of the event. The laziness arises from
the too much work that needs to be done in order to learn exceptionless rules
and it is too hard to learn such rules. The theoretical and practical ignorance
arises from lack of complete evidence and it is not possible to conduct all the
necessary tests to learn a particular event. Hence, the domain knowledge, and
in our case the domain conceptualization, can at best provide only a degree of
belief of the relevant groups of words. We use probability theory to deal with the
degrees of belief. As mentioned in [
          <xref ref-type="bibr" rid="ref19">19</xref>
          ], the probability theory has the same
ontological commitment as the formal logic, though the epistemological commitment
di ers. The process of learning and presenting a probabilistic conceptualization
is divided into four phases as shown in Figure 1. They are, 1) pre-processing
2) syntactic analysis 3) semantic analysis, and 4) representation.
3.1
        </p>
        <p>
          Pre-processing
A corpus contains a plethora of structured and unstructured sentences. A lexicon
of a language is its vocabulary built from lexemes [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ], [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ]. A lexicon contains
words belonging to a language and in our work individual words from the corpus.
In pure form, the lexicon may contain words that appear frequently in the corpus
but have little value in formalizing a meaningful criterion. These words are called
stop words or in our terminology: negated lexicon, and they are excluded from the
vocabulary. We, rst, part-of-speech tagged the corpus with the Penn Treebank
English POS tag set [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ]. We use the subset of tagset NN, NNP, NNS, NNPS, JJ,
JJR, JJS, VB, VBD, VBG, VBN, VBP, and VBZ. The word length WL above
some threshold WLT is also considered. The length of a word, with respect to
POS context, is the sequence of characters or symbols that made up the word.
By default, we consider that a word with WL &gt; 2 su ciently formalizes to some
criterion.
        </p>
        <p>
          The pure form of the lexicon might contain words that need to be further
puri ed according to some criterion. We use regular expressions for this task.
Then we normalize and case-fold the words [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ]. In addition to this there are
families of derivationally related words with similar meanings. We use stemming
and lemmatization to reduce the in ectional forms and derivational forms of a
word to a common base form [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ]. We achieve this with the aid of WordNets'
stemming algorithms. We couple the knowledge of POS tag of the word to get
the correct context when nding the common base form.
3.2
        </p>
        <p>Syntactic Analysis
The primary focus on this phase is to look at the structure of the sentences and
learn the associations among the vocabulary. We assume that each sentence of
the corpus follows the POS pattern 1. 1,
(SubjectNounP hrase+)(V erb+)(ObjectNounP hrase+)
(1)</p>
        <p>
          We hypothesize that the associations learned from this phase provides the
potential candidates for concepts and relations of the ontology. But the vocabulary
itself does not provide su cient ontology concepts. We use a notion of grouping
of consecutive sequence of words to form an OWL concept. This grouping is done
using an appropriate N-gram model [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ]. We illustrate this idea using Figure 2.
        </p>
        <p>The group w1 w2 forms a potential concept in the conceptualization. We
use the notation x y to show that the word y is appended to the word x. The
groups w2 w3, w3 w4 etc. form other potential concepts in the conceptualization.
Word w3 comes after group w1 w2. According to the Bayes viewpoint, we collect
information to estimate the probability P (w3jfw1 w2g), which will be used to
form IS-A relationships, w1 w2 v w3 using an independent Bayesian network
with conditional probability P (fw1 w2gjw3). In addition to this, we count the
groups appear in the left hand side and the right hand side of the expression 1
and the association of of these groups given the verbs. These counts are used in
the third phase to create the relations among concepts.
This phase conducts the semantic analysis with probabilistic reasoning, which
constitutes the most important operation of our work. This phase determines the
conceptualization of the domain using a probability distribution for IS-A
relations and relations among the concepts. Our main de nition of concept learning
is given in De nition 1.</p>
        <p>De nition 1. The set W = fw1; : : : ; wng represents words of the vocabulary
and each wi has a prior probability i &gt; . is a prior threshold, which is known
as the knowlege factor. The set G = fg1; : : : ; gmg represents N-gram groups
learned from the corpus and each gj has a prior probability j . When w 2 W
and g 2 G, P (wjg) is the likelihood probability learned from the corpus. The
entities w and g represent the potential concepts of the conceptualization and
the set W provide the potential super-concepts of the conceptualization. Within
this environment, an IS-A relationship between w and g is given by the posterior
probability P (gjw) and this is represented with a Bayesian network having two
nodes w and g and is modeled by the equation,</p>
        <p>P (gjw) =</p>
        <p>Pi p(wjgi)
p(gi)
:
(2)</p>
        <p>Using the De nition 1, the probabilistic conceptualization of a domain is
de ned as follows.</p>
        <p>De nition 2. The probabilistic conceptualization of the domain is represented
by an n-number of independent Bayesian networks sharing groups.</p>
        <p>Figure 3 shows a simple example of the De nition 2. The interpretation of
De nition 2 is: Let a set G contains an n-number of nite random variables
fg1; : : : ; gng. There exist a group gi, which is shared by m words fw1; : : : ; wmg.
Then, with respect to the Bayesian framework, BNi of P (gijwi) is calculated and
max(P (gijmi)) is selected for the construction of the ontology. This means that if
there exists two Bayesian networks and the Bayesian network one is given by the
pair w1; g1 and the Bayesian network two is given by the pair fw2; g1g then the
Bayesian network that has the most substantial IS-A relationship is obtained
through maxBNi (P (g1jw1)) and this network is retained and other Bayesian
networks will be ignored when building the ontology. If all P (g1jw1) remains
equal, then the Bayesian network with the highest super-concept probability
will be retained. These two conditions will resolve any naming issues.</p>
        <p>The next step is to induce the relationships to complete the
conceptualization. In order to do this, we need to nd semantics associated with each verb.
We hypothesize that relations are generated by the verbs and the de nition is
as follows.</p>
        <p>De nition 3. The relationships of the conceptualization are learned from the
syntactic structure model by the expression 1 and the semantic structure model
by the lambda expression obj: sub:V erb(sub; obj), where -reduction is applied
for obj and sub of the expression 1. If there exists a verb V between two groups
of concepts C1 and C2, the relationship of the triple (V; C1; C2) is written as
V (C1; C2) and model with conditional probability P (C1; C2jV ). The Bayesian
network for relationship is and the model semantic relationship is given by,</p>
        <p>P (C1; C2jV ) = p(C1jV )p(C2jV ) ! V (C1; C2)</p>
        <p>The relations learned from De ntions 3 needs to be subjected to a lower
bound. The lower bound is known as the relations factor. When the corpus is
substantially large, the number of relations is proportional to the number of
verbs. Not all relations may relevant and the factor is used as the threashold. A
verb may have antonyms. If a verb is associated with some concepts and these
concepts happen to be associated with a antonym, the verb with the highest
Bayesian probability value is selected for the relations map and the other
relationship will be removed. Finally, the probabilistic conceptualization is serialized
as an OWL DL ontology in the representation phase.</p>
        <p>Our implementation of the above phases is based on Java 6 and it is named
as PrOntoLearn (Probabilistic Ontology Learning).
4</p>
        <sec id="sec-4-3-1">
          <title>Experiments</title>
          <p>We have conducted experiments on three main data corpora, 1) the PCAssay, of
the BioAssay Ontology (BAO) project, Department of Molecular and Cellular
Pharmacology University of Miami, School of Medicine 2) a sample collection of
38 PDF les from ISWC 2009 proceedings, and 3) a substantial portion of the
web pages extracted from the University of Miami, Department of Computer
Science7 domain . We have constructed ontologies for all three corpora with
di erent parameter settings.</p>
          <p>The rst corpus contains high throughput screening assays performed on
various screening centers. This corpus grows rapidly each month. We speci
cally limited our dataset to assays available on the 1st of January 2010. Table
2 provides the statistics of the corpus. We extract the vocabulary generated</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>7 http://www.cs.miami.edu</title>
      <p>from [a-zA-Z]+[- ]? nw* regular expression, and normalized them to create the
vocabulary.</p>
      <p>The average le size of the corpus is approximately 6 Kb. We conducted
these experiments in a Genuine Intel(R) CPU 585 @ 2.16GHz, 32 bits, 2 Gb
Toshiba laptop. It is found that the time required to build the
conceptualization grows linearly. We use precision, recall and F1 measures to evaluate the
ontology and recommendations from domain experts, specially to get comments
on the generated bioassay ontology. The ontology that is generated is too large
to show in here.Instead, we provide a few distinct snapshots of the ontology
with the help of Protege OWLViz plugin. Figures 5 and 6 show snapshots of
the ontology created from the BioAssay Ontology corpus for input parameters
KF = 0:5, N-gram = 3; and RF = 0:9. Figure 5 shows the IS-A relationships
and Figure 6 shows the binary relationships.</p>
      <p>
        According to experts, the ontology contains rich set of vocabulary, which is
very useful for top-down ontology construction. The experts also mentioned that
the ontology has good enough structure. The www:cs:miami:edu corpus is used
to calculate quantitative measurements. The gold standard based approaches
such as precision (P ) and recall (R) and F-measure (F1) are used to evaluate
ontologies [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. We use a slightly modi ed version of [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ] as our reference ontology.
Table 3 shows the results. The average precision of the constructed ontology is
approximately 42%. It is to be noted that we use only one reference ontology. If
we use another reference ontology the precision values varies. This means that
the precision value depends on the available ground truth.
      </p>
      <p>The results show that our method creates an ontology for any given domain
with acceptable results. This is shown in the precision value, if the ground truth
is available. On the other hand, if the domain does not have ground truth the
results are subject to domain expert evaluation of the ontology. One of the
potential problems we have seen in our approach is search space. Since our method
is unsupervised, it tends to search the entire space for results, which is
computationally costly. We thus need a better method to prune the search space so that
out method provide better results. According to domain experts, our method
extracts good vocabulary but provides a at structure. They have proposed a
sort of a semi-supervised approach to correct this problem, by combining the
knowledge from domain experts and results produced by our system. We left the
detailed investigation for future work.</p>
      <p>Since our method is based on the Bayesian reasoning (which uses N-gram
probabilities), it is paramount that the corpus contains enough evidence of the
redundant information. This condition requires that the corpus to be large enough
so that we can hypothesize that the corpus provides enough evidence to build
the ontology.</p>
      <p>We hypothesize that a sentence of the corpus would generally be subjected to
the grammar rule given in expression 1. This constituent is the main factor that
uses to build the relationships among concepts. In NLP, there are many other
ner grained grammar rules that speci cally t for given sentences. If these
grammar rules are used, we believe we can build a better relationship model.
We have left this for future work.</p>
      <p>
        At the moment our system does not distinguish between concepts and the
individuals of the concepts. The learned A-Box primarily consists of the
probabilities of each concepts. This is one area where we are eager to work on. Using
the state-of-the art NLP techniques, we plan to ll this gap in a future work.
Since our method has the potential to be used in any corpus, it could be seen
that the lemmatizing and stemming algorithms that are available in WordNet
would not recognize some of the words. Specially in the BioAssay corpus, we
observe that some of the domain speci c words are not recognized by WordNet.
We use the Porter stemming algorithm [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ] to get the word form and it shows
that this algorithm constructs peculiar word forms. Therefore, we deliberately
remove it from the processing pipeline.
      </p>
      <p>The complexity of our algorithms is as follows. The bootstrapping algorithm
available in the syntactic layer has a worst case running time of O(M max(sj )
max(wk)), where M is the number of documents, sj is a the number of sentences
in a document, and wk is the number of words in a sentence. The probabilistic
reasoning algorithm has the worst case running time of O(jLj jSuperConceptsj),
where jLj is the size of the lexicon and jSuperConceptsj is the size of the
super concepts set. The ontologies generated from the system are consistent with
Pellet8 and FaCT++9 reasoners.</p>
      <p>Finally, our method provides a process to create a lexico-semantic ontology
for any domain. For our knowledge, this is a very rst research on this line of</p>
    </sec>
    <sec id="sec-6">
      <title>8 http://clarkparsia.com/pellet</title>
    </sec>
    <sec id="sec-7">
      <title>9 http://owl.man.ac.uk/factplusplus/</title>
      <p>work. So we continue our research along this line and to provide better results
for future use.
5</p>
      <sec id="sec-7-1">
        <title>Conclusion</title>
        <p>We have introduced a novel process to generate an ontology for any random text
corpus. We have shown that our process constructs a exible ontology. It is also
shown that in order to achieve high precision, it is paramount that the corpus
should be large enough to extract important evidence. Our research has also
shown that probabilistic reasoning on lexico-semantic structures is a powerful
solution to overcome or at least mitigate the knowledge acquisition bottleneck.
Our method also provides evidence to domain experts to build ontologies
using a top-down approach. Though we have introduced a powerful technique to
construct ontologies, we believe that there is a lot of work that can be done to
improve the performance of our system. One of the areas our method lacks is the
separation between concepts and individuals. We would like to use the generated
ontology as a seed ontology to generate instances for the concepts and extract
the individuals already classi ed as concepts. Finally, we would like to increase
the lexicon of the system with more tags available from the Penn Treebank tag
set. We believe that if we introduce more tags into the system, our system can
be trained to construct human readable (friendly) concepts and relations names.</p>
      </sec>
      <sec id="sec-7-2">
        <title>Acknowledgements</title>
        <p>This work was partially funded by the NIH grant RC2 HG005668.</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Banerjee</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pedersen</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          :
          <article-title>The design, implementation and use of the n-gram statistics package</article-title>
          .
          <source>In: In Proceedings of the Fourth International Conference on Intelligent Text Processing and Computational Linguistics</source>
          . pp.
          <volume>370</volume>
          {
          <issue>381</issue>
          (
          <year>2003</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Bos</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Markert</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          :
          <article-title>Recognising textual entailment with logical inference</article-title>
          .
          <source>In: HLT '05: Proceedings of the conference on Human Language Technology and Empirical Methods in Natural Language Processing</source>
          . pp.
          <volume>628</volume>
          {
          <fpage>635</fpage>
          . Association for Computational Linguistics, Morristown, NJ, USA (
          <year>2005</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Carlson</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Betteridge</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>R.C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hruschka</surname>
          </string-name>
          , Jr., E.R., Mitchell, T.M.:
          <article-title>Coupled semi-supervised learning for information extraction</article-title>
          .
          <source>In: WSDM '10: Proceedings of the third ACM international conference on Web search and data mining</source>
          . pp.
          <volume>101</volume>
          {
          <fpage>110</fpage>
          .
          <string-name>
            <surname>ACM</surname>
          </string-name>
          , New York, NY, USA (
          <year>2010</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Chemudugunta</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Holloway</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Smyth</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Steyvers</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Modeling documents by combining semantic concepts with unsupervised statistical learning</article-title>
          .
          <source>In: ISWC '08: Proceedings of the 7th International Conference on The Semantic Web</source>
          . pp.
          <volume>229</volume>
          {
          <fpage>244</fpage>
          . Springer-Verlag, Berlin, Heidelberg (
          <year>2008</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Cimiano</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hotho</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Staab</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Learning concept hierarchies from text corpora using formal concept analysis</article-title>
          .
          <source>Journal of Arti cial Intelligence research 24</source>
          ,
          <volume>305</volume>
          {
          <fpage>339</fpage>
          (
          <year>2005</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Cimiano</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          , Volker, J.:
          <article-title>Text2Onto - A Framework for Ontology Learning and Data-driven Change Discovery (</article-title>
          <year>2005</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Clark</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Harrison</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Large-scale extraction and use of knowledge from text</article-title>
          .
          <source>In: K-CAP '09: Proceedings of the fth international conference on Knowledge capture</source>
          . pp.
          <volume>153</volume>
          {
          <fpage>160</fpage>
          .
          <string-name>
            <surname>ACM</surname>
          </string-name>
          , New York, NY, USA (
          <year>2009</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Dellschaft</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Staab</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Strategies for the evaluation of ontology learning</article-title>
          .
          <source>In: Proceeding of the 2008 conference on Ontology Learning and Population: Bridging the Gap between Text and Knowledge</source>
          . pp.
          <volume>253</volume>
          {
          <fpage>272</fpage>
          . IOS Press, Amsterdam, The Netherlands, The
          <string-name>
            <surname>Netherlands</surname>
          </string-name>
          (
          <year>2008</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Ding</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Peng</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          :
          <article-title>A Probabilistic Extension to Ontology Language OWL</article-title>
          .
          <source>In: HICSS '04: Proceedings of the Proceedings of the 37th Annual Hawaii International Conference on System Sciences (HICSS'04) - Track 4</source>
          . p.
          <volume>40111</volume>
          .
          <article-title>1</article-title>
          . IEEE Computer Society, Washington, DC, USA (
          <year>2004</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Gruber</surname>
            ,
            <given-names>T.R.:</given-names>
          </string-name>
          <article-title>A translation approach to portable ontology speci cations</article-title>
          .
          <source>Knowledge Acquisition</source>
          <volume>5</volume>
          (
          <issue>2</issue>
          ),
          <volume>199</volume>
          {
          <fpage>220</fpage>
          (
          <year>1993</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Jurafsky</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Martin</surname>
            ,
            <given-names>J.H.</given-names>
          </string-name>
          :
          <article-title>Speech and Language Processing: An Introduction to Natural Language Processing, Computational Linguistics and Speech Recognition</article-title>
          . Prentice Hall, Pearson Education International,
          <volume>2</volume>
          . edn. (
          <year>2009</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Koller</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Levy</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pfe</surname>
            <given-names>er</given-names>
          </string-name>
          , A.:
          <string-name>
            <surname>P-CLASSIC</surname>
          </string-name>
          :
          <article-title>A tractable probabilistic description logic</article-title>
          .
          <source>In: Proceedings of AAAI-97</source>
          . pp.
          <volume>390</volume>
          {
          <issue>397</issue>
          (
          <year>1997</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Lin</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pantel</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Discovery of inference rules for question-answering</article-title>
          .
          <source>Natural Language Engineering</source>
          <volume>7</volume>
          (
          <issue>4</issue>
          ),
          <volume>343</volume>
          {
          <fpage>360</fpage>
          (
          <year>2001</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Lukasiewicz</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          :
          <article-title>Probabilistic description logics for the semantic web</article-title>
          .
          <source>Tech. rep., Nr</source>
          .
          <fpage>1843</fpage>
          -
          <volume>06</volume>
          -05, Institut fur Informationssysteme, Technische Universitat Wien (
          <year>2007</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Manning</surname>
            ,
            <given-names>C.D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Raghavan</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          , Schutze, H.: Introduction to Information Retrieval. Cambridge University Press, New York, NY, USA (
          <year>2008</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <surname>Marcus</surname>
            ,
            <given-names>M.P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Marcinkiewicz</surname>
            ,
            <given-names>M.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Santorini</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          :
          <article-title>Building a large annotated corpus of english: the penn treebank</article-title>
          .
          <source>Comput. Linguist</source>
          .
          <volume>19</volume>
          (
          <issue>2</issue>
          ),
          <volume>313</volume>
          {
          <fpage>330</fpage>
          (
          <year>1993</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          17.
          <string-name>
            <surname>Poon</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Domingos</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          : In:
          <article-title>Proceedings of the Forty-Eighth Annual Meeting of the Association for Computational Linguistics</article-title>
          , Uppsala, Sweden.
          <source>ACL</source>
          (
          <year>2010</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          18.
          <string-name>
            <surname>Porter</surname>
            ,
            <given-names>M.F.</given-names>
          </string-name>
          :
          <article-title>An algorithm for su x stripping</article-title>
          .
          <source>Program</source>
          <volume>14</volume>
          (
          <issue>3</issue>
          ),
          <volume>130</volume>
          {
          <fpage>137</fpage>
          (
          <year>1980</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          19.
          <string-name>
            <surname>Russell</surname>
            ,
            <given-names>S.J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Norvig</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Arti cial Intelligence: A Modern Approach</article-title>
          . Prentice Hall,
          <volume>3rd</volume>
          <fpage>edn</fpage>
          . (
          <year>2009</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          20.
          <string-name>
            <surname>Salloum</surname>
            ,
            <given-names>W.:</given-names>
          </string-name>
          <article-title>A question answering system based on conceptual graph formalism</article-title>
          .
          <source>In: KAM '09: Proceedings of the 2009 Second International Symposium on Knowledge Acquisition and Modeling</source>
          . pp.
          <volume>383</volume>
          {
          <fpage>386</fpage>
          . IEEE Computer Society, Washington, DC, USA (
          <year>2009</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          21. SHOE:
          <article-title>Example computer science department ontology</article-title>
          , http://www.cs.umd.edu/ projects/plus/SHOE/cs.html,
          <source>last visited on June 14</source>
          ,
          <year>2010</year>
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          22.
          <string-name>
            <surname>Studer</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Benjamins</surname>
            ,
            <given-names>V.R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fensel</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>Knowledge engineering: Principles and methods</article-title>
          .
          <source>Data and Knowledge Engineering</source>
          <volume>25</volume>
          (
          <issue>1-2</issue>
          ),
          <volume>161</volume>
          {
          <fpage>197</fpage>
          (
          <year>1998</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          23.
          <string-name>
            <surname>Tanev</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Magnini</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          :
          <article-title>Weakly supervised approaches for ontology population</article-title>
          .
          <source>In: Proceeding of the 2008 conference on Ontology Learning and Population: Bridging the Gap between Text and Knowledge</source>
          . pp.
          <volume>129</volume>
          {
          <fpage>143</fpage>
          . IOS Press, Amsterdam, The Netherlands (
          <year>2008</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          24. Volker, J.,
          <string-name>
            <surname>Hitzler</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cimiano</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Acquisition of owl dl axioms from lexical resources</article-title>
          .
          <source>In: ESWC '07: Proceedings of the 4th European conference on The Semantic Web</source>
          . pp.
          <volume>670</volume>
          {
          <fpage>685</fpage>
          . Springer-Verlag, Berlin, Heidelberg (
          <year>2007</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>