=Paper=
{{Paper
|id=Vol-289/paper-3
|storemode=property
|title=Knowledge Capturing Tools for Domain Experts
|pdfUrl=https://ceur-ws.org/Vol-289/p03.pdf
|volume=Vol-289
|dblpUrl=https://dblp.org/rec/conf/kcap/BrockerRW07
}}
==Knowledge Capturing Tools for Domain Experts==
Knowledge Capturing Tools for Domain Experts
Exploiting Named Entity Recognition and n-ary Relation Discovery for Knowledge
Capturing in E-Science
Lars Bröcker Marc Rössler Andreas Wagner
Fraunhofer IAIS Computational Linguistics Computational Linguistics
Schloss Birlinghoven University of Duisburg-Essen University of Duisburg-Essen
53754 Sankt Augustin, 47048 Duisburg, Germany 47048 Duisburg, Germany
Germany marc.roessler@uni- andreas.wagner@uni-
lars.broecker@iais.fhg.de due.de due.de
ABSTRACT gence]: Learning—knowledge acquisition, concept learning;
The success of the Semantic Web depends on the availabil- I.2.7 [Artificial Intelligence]: Natural Language Process-
ity of content marked up using its description languages. ing—text analysis; I.5.3 [Pattern Recognition]: Cluster-
Although the idea has been around for nearly a decade, ing
the amount of Semantic Web content available is still fairly
small. This is despite the existence of many digital archives General Terms
containing lots of high quality collections which would, ap- Algorithms
propriately marked up, greatly enhance the reach of the Se-
mantic Web. The archives themselves would benefit as well, Keywords
by improved opportunities for semantic search, navigation
Named Entity Recognition, Relation Discovery, Semantic
and interconnection with other archives.
Networks, Wiki Systems
The main challenge lies in the fact that ontology creation at
the moment is a very detailed and complicated process. It 1. INTRODUCTION
mostly requires the service of an ontology engineer, who de- The Semantic Web can only flourish if enough content provi-
signs the ontology in accordance with domain experts. The ders adopt it for the presentation of their content. This lack
software tools available, be it from the text engineering or of adoption is the Achilles heel of the vision of the data web
the ontology creation disciplines, reflect this: they are built where humans and software agents can work side by side.
for engineers, not for domain experts. In order to really tap The main reason for this lies right at the base of the Semantic
the potential of the digital collections, tools are needed that Web: the creation of ontologies. The process needed to get
support the domain experts in marking up the content they to a working representation of a domain is too difficult for
understand better than anyone else. domain experts to do it on their own - a debilitating factor
on the way to widespread adoption: the WWW did flourish
This paper presents an integrated approach to knowledge simply due to the ease of marking up knowledge in HTML.
capturing and subsequent ontology creation, called WIKIN- This does not hold true for OWL or even RDF.
GER, that aims at empowering domain experts to prepare
their content for inclusion into the Semantic Web. This is There are tools that deliver support in the process of creat-
done by largely automating the process through the use of ing an ontology, both from the domain of text engineering
named entity recognition and relation discovery. as well as from ontology engineering. But these tools are
made for a selected audience: ontology engineers. This in
itself is nothing bad, but it reduces the amount of growth of
Categories and Subject Descriptors the Semantic Web to the availability (and affordability) of
H.3.1 [Information Storage and Retrieval]: Content said engineers. Tools are needed that allow domain experts
Analysis and Indexing—linguistic processing; I.2.4 [Artifi- themselves to design and create ontologies tailored for their
cial Intelligence]: Knowledge Representation Formalisms needs and domain corpora, if the Semantic Web is to come
and Methods—semantic networks; I.2.6 [Artificial Intelli- about on a grand scale.
But what is needed to create an ontology from a text corpus?
First of all, an ontology can be seen as a graph structure,
a semantic network. The nodes of this graph are the enti-
ties, i.e. the actors, topics and objects of the ontology, while
the edges of the graph are the relations that exist between
the entities. The task of automatically creating an ontology
can be broken down into the following steps: first named en-
tity recognition (NER) and second the detection of relations
existing between those entities.
The detection and classification of proper names into prede- The architecture of WIKINGER is motivated by the as-
fined categories is called Named Entity Recognition (NER). sumption that many nodes of a domain specific semantic
The recognition of the categories PERSON, LOCATION network occur in domain relevant texts and that these oc-
and ORGANIZATION within the newspaper domain is es- currences are proper names or expressions which can be ex-
pecially well-studied as a part of the MUC-campaigns (Mes- tracted with NER-techniques.
sage Understanding Conferences) and can be conducted au-
tomatically with a performance beyond 0.9 F-measure for The pilot domain of WIKINGER is contemporary history
English texts [4]. The detection of relations between the en- with a focus on the history of Catholicism in Germany. For
tities of a corpus is a younger discipline, usually concerned that domain, the traditional NER categories PERSON, LO-
with binary relations. Experiments on English newspapers CATION, ORGANIZATION, and TIME/DATE expressions
show performance around 0.75 F-measure [8]. These ad- obviously carry crucial nodes for a domain specific seman-
vances facilitate a largely automated processing of text cor- tic network. However, the domain experts desired additional
pora into domain ontologies. This paper introduces an in- categories, such as HISTORICAL-EVENT, BIOGRAPHIC-
tegrated web service-based framework called WIKINGER EVENT or ROLE. A ROLE is a function or a position a per-
that does just that. son holds (e.g. ”bishop”, ”professor of theology”) and is often
part of a BIOGRAPHIC-EVENT, which may contain addi-
This paper is structured as follows: Section 2 gives an o- tional annotations such as LOCATION and TIME/DATE,
verview of the WIKINGER framework, sections 3 and 4 de- as the following example shows:
scribe our work on named entity extraction, while section 5
describes the relation discovery part of the process. After
that, section 6 highlights relevant related work, and we close
with remarks on future works and the conclusion in sections 1936
7 and 8. archbishop of
Cologne
2. WIKINGER - THE BIG PICTURE
WIKINGER[3], short for Wiki Next Generation Enhanced The HISTORICAL-EVENT describes events significant to
Repositories, aims at developing collaborative knowledge plat- the domain experts, such as the ”Wall Street Crash of 1929”,
forms for scientific communities. The collaboration is fa- also called ”Black Thursday”. This category may contain
cilitated by selecting a Wiki as a presentation layer, and embedded categories, too. The two event categories of the
the knowledge contained can be organized via semantic re- pilot domain are beyond the traditional NER task: Depend-
lations. The resulting semantic Wiki can be extended, re- ing on the perspective, they either involve relation extraction
organized and commented on by all (registered) members of or embedded categories. The corpus to annotate currently
the particular scientific community. To setup and maintain consists of approximately 150 monographs within a book
the semantic network, NER-techniques are applied to the series. The books were scanned and the text was OCR-
available domain-relevant documents (see section 3). The extracted. The annotations of the resulting corpus will be
resulting annotations are the potential nodes of the seman- used as potential nodes of the semantic network to be cre-
tic network that is constructed in a semi-automatic manner. ated.
The relations are proposed based on clusters of co-occurring
entities (see section 5). Since the book series has a consistent layout structure, it
was possible to preserve some layout information, such as
Figure 1 shows a view of the components that are part of the distinction between footnotes and other text. This dis-
the WIKINGER framework. It is built following a service- tinction is helpful in order to detect a text unit specific to
oriented architecture, its modules are loosely coupled, which the texts of our domain called a ”biogram”. A biogram usu-
allows need-driven reconfiguration of the system. The sys- ally is a footnote that is provided the first time a person
tem itself uses a linked set of data repositories to perform is mentioned in the text and comprises a short biography.
its duties. The resource layer at the bottom of fig. 1 shows These biographies usually are short and concise and tend
a drastically simplified view of the outside world: it con- to follow a predetermined structure. For instance, most of
tains arbitrary data sources that can be imported into the the biograms start with the name of the person, and some
first of the repositories, i.e. the document repository. This biograms present the single pieces of information separated
repository provides the other services of the system with a by a particular delimiter such as semicolon or comma.
versioned corpus of documents to work on. The process-
ing services (e.g. for NER, relation discovery and creation Thus, in most cases the person named at the beginning of
of the ontology) use this repository as a source only. They a biogram is the one that the other annotations in that bi-
feed their results into the metadata repository. It is linked ogram relate to. While some of the information items also
to the document repository to uphold references to the orig- belong to persons that are related to the person described in
inal and it also provides versioned storage of the data. This the biogram (e.g. ”his father was a prime minis-
ensures that the original corpus remains unchanged. The ter ”) this assumption nevertheless holds true for
final repository contains the semantic model of the corpus. the largest part of the corpus. This is very important for the
It makes use of both the document repository as well as the relation discovery step, since all relations discovered in a spe-
metadata repository. At the moment, the application layer cific biogram are linked implicitly to said person, although
takes the form of a wiki system, but other applications can its participation in most of the relations is not readily appar-
easily be envisioned. ent from their local contexts. Accordingly, they need to be
WIKINGER System Components
Application Layer
Guest/Member Author/Editor Annotator Administrator
Browser Editor Annotation Administration
Entity Model Processing Service
Entity Service
Repository
Service Layer
Analyzer
Meta Data NER
Meta Data Service ...
Repository
Document Document Service Account Service
Repository
Resource Layer
User
Document
Database
Sources
Figure 1: The WIKINGER Framework: Component View
associated with the person discussed in the biogram, which A well-known example for such a task is the recognition of
in turn has implications for the creation of the semantic net- biomedical entities such as genes, proteins or cell tissue [6,
work from the anntoation and relation data discovered in the 9]. It is almost impossible for a non-expert in the biomedical
course of the process. domain to judge about the correctness of an annotation or
even to figure out a definition of the classes to recognize.
Processing these biograms results in a semantic network in Additionally, capitalization is not a distinctive feature of
OWL which contains any information that could be har- the entities to detect. Furthermore, biomedical entities are
vested automatically from all the biograms within the 150 no proper names in the linguistic sense since a mention of a
monographs. This knowledge base constitutes a biographi- particular protein refers to all instances of that protein and
cal database for the scientific domain, which, according to not to a particular instance.
the historians working within the WIKINGER project, is
a long time desideratum for the domain of contemporary The annotation task within WIKINGER has similar char-
history of Catholics in Germany. acteristics: the documents to be processed are specialized
texts, thus the definition of the annotation categories has to
However, the tasks described are not limited to the pilot be provided by the domain experts. Also, most of the texts
application of WIKINGER. Indeed, it has many features in are in German, so the capitalization is not a reliable clue to
common with a series of annotation tasks found in other do- detect proper names. Furthermore, discussions with the do-
mains as well. Our research within the WIKINGER project main experts have shown that some of the annotation tasks
focuses on the application-oriented generalization of these amount to information extraction in a more general sense,
challenges. in particular involving relation extraction, even though on
a local level. For example, the BIO-EVENT provided in
section 2 establishes a relation between the person the re-
3. NER spective biogram deals with, a role occupied by that person,
It is highly desirable to generalize successful NER approa- a certain time, and a location. Although these annotation
ches described in section 1 to a broader variety of semantic tasks significantly expand the annotation of proper names,
markup at phrase level (i.e. apart from ”standard” categories we still consider them as a sophisticated form of NER. In
such as PERSON, ORGANIZATION, or LOCATION) in other words, we basically employ approaches which have
order to support other NLP applications. However, this re- been successfully applied to NER.
quires annotation components that can be extended to new
categories and adapted to new domains and new languages. In principle, two major kinds of NER approaches have been
These tasks may have different characteristics than the clas- proposed in the literature: rule-based and machine learning
sical MUC task: First, they may lack the clue of the dis- (ML) approaches. Rule-based approaches employ a hand-
tinctive capitalization for some semantic classes and some crafted set of rules which is fine-tuned to the particular ap-
languages, such as German. Second, the categories of inter- plication domain. The adaptation of such a rather complex
est may neither be obvious nor easily understandable due to rule set to new domains and/or languages brings about ex-
a highly specialized domain and language.
tensive modification and maintenance efforts and requires 4. WALU
therefore comprehensive knowledge about both the new do- A prerequisite for enabling domain experts to create training
main and the proper design of the linguistic rule set. This data and control the process of training and (semi-)automa-
means that domain experts need extensive support by com- tic semantic markup is the availability of a powerful and
putational linguists in order to port such a system to their convenient tool. On the one hand, such a tool has to pro-
domain. In contrast, adapting machine learning approaches vide the necessary functionalities, i.e. manual annotation
to a new application domain requires the creation of domain- of documents, configuration and initiation of the training
specific training data, i.e. manual annotation of domain- process, application of automatic annotation components,
specific documents. Since this essentially requires domain as well as inspection and correction of the resulting anno-
(rather than linguistic) expertise, domain professionals need tations. On the other hand, intuitive interfaces and con-
much less support by computational linguists (if any at all). venient facilities supporting these functionalities while en-
Our experience within the WIKINGER project has shown capsulating their complexity are crucial to ensure usability
that such support is necessary primarily for the initial task for professionals of any domain. In addition, this tool has
of defining a suitable set of semantic categories. During this to be integrated into the overall WIKINGER infrastructure
definition stage, the communication between domain experts sketched in section 2. Currently there is no tool available
and linguists in essence consists in exchanging annotated that meets all these requirements (see section 6), at least
examples. We believe that this example-based communica- not to our knowledge. Therefore, we are developing such a
tion significantly facilitates portability, since concrete ex- tool, which we call WALU (WIKINGER Annotations- und
amples are much easier to create and understand than the Lern-Umgebung = WIKINGER annotation and learning en-
explicit formulation of more or less complex and abstract vironment, see [16]).
(sub-)regularities. The same holds true for the annotation
of the training data itself, which can be regarded as example- WALU supports manual annotation with a GUI that is easy
based communication between domain experts and machine to use. It offers a comfortable navigation through the an-
learning algorithms. notations, and simple but effective annotation support such
as the automatic adjustment of markup boundaries or a dy-
Consequently, in order to minimize the amount of “external namic markup dictionary. This dictionary is created during
help” specialists needed to set up the WIKINGER system the annotation process and is used to propose markup la-
for their domain, we decided to employ ML approaches for bels for text passages corresponding to dictionary entries.
NER. In our current experiments, we are using Maximum Using a context-sensitive menu, the annotator confirms or
Entropy modeling and support vector machines. (As im- rejects these proposals and/or removes the entry from the
plementations, we employ openNLP1 and SVMstruct2 , re- dictionary. In our experience the immediate feedback of the
spectively.) However, we aim at providing a variety of ML dynamic markup dictionary also helps the domain experts
algorithms which can either be employed independently or in to clarify the task of string-based identification of domain-
combination to maximize performance. Regarding portabil- relevant concepts. Additionally, WALU also provides an au-
ity, it is crucial that the learning approaches employ domain- tomatic annotator for strings referring to the category DATE
independent features and resources that can be easily adap- which is based on regular expressions. This is a simple pro-
ted to a new domain or a new NER task. Furthermore, these totype of a series of automatic mechanisms that will be used
methods have to be applied in a way that allows the acqui- to annotate all the available documents. Except a few anno-
sition of embedded annotations. “Standard” ML classifiers tators based on regular expressions to classify entities with
assign one class (in our case, a semantic category) to each unique patterns (such as email addresses and URLs), most of
instance to classify (in our case, a token)3 . In embedded these annotators are based on machine learning algorithms
annotations, (parts of) entities may receive multiple classes that will be accessible via WALU.
simultaneously (e.g. in the example in section 2, “1936” is
at the same time a DATE and part of a BIO-EVENT). To Training the ML facilities mentioned in section 3 as well as
achieve such kind of concurrent classification, we run multi- their annotation of new text can be initiated via the WALU
ple classifiers, each one assigning different classes, and unify GUI. The annotation results can be displayed and manually
the results. For ML approaches which are restricted to bi- corrected. Automatic annotations are displayed in a distinct
nary classification (e.g. SVM), one classifier is required for way (only the lower half of the annotated tokens are marked)
each category. For ML approaches without this restriction so that they can be discovered immediately by the user.
(e.g. MaxEnt), classifiers assigning multiple classes can be
built and combined in a more flexible way. Our experiments WALU is designed both as a part of the WIKINGER infras-
with MaxEnt models have shown that combining classifiers tructure and as a stand-alone tool. Web-service-based com-
each of which assigns all categories except one, i.e. each munication facilities allow WALU to load documents from
of which “ignores” one particular class, yields higher perfor- the WIKINGER document repository and load/store corre-
mance than employing binary classifiers. In these experi- sponding annotations from/to the metadata repository. As
ments, we got F-measures (at token level) of up to 84.6% a stand-alone tool, WALU currently is able to import text
for persons, 87,1% for organizations, 94,8% for geographic- documents (other import formats will be captured later) and
political entities, and 92,8% for roles. to export annotated documents in a straightforward XML
standoff format. The transfer between the various different
1 data formats is achieved via a special internal format we
http://maxent.sourceforge.net/ call ‘WaRP (WALU Rich Paragraph) stream’, which is also
2
http://svmlight.joachims.org/svm struct.html processed by the automatic annotation components.
3
Multiword NEs are recognized as a sequence of tokens re-
ceiving the same class.
relation. Since the amount of relation clusters is not known
beforehand, agglomerative clustering is applied. In this al-
gorithm, every vector starts as its own cluster. Clusters
are then merged, given they fulfill a certain clustering crite-
rion that is defined on a distance measure. We use standard
Cosine similarity as distance and allow both single and com-
plete linkage as criteria. Given two clusters A and B and a
distance threshold t, this translates to:
Figure 2: Workflow of the algorithm
Single Linkage : ∃α ∈ A, β ∈ B : min(dist(α, β)) < t
5. SEMIAUTOMATIC RELATION DISCO-
Complete Linkage : ∃α ∈ A, β ∈ B : max(dist(α, β)) < t
VERY
The algorithms and tools described in the preceding sections
provide named entities for a variety of project-dependent Which method will be used depends on the corpus in ques-
concept classes. They will become the nodes of the semantic tion. Terse texts show better results with complete linkage,
network that is to be built. The remaining part is the provi- normal text performs better with single linkage.
sion of edges connecting these nodes, which will be explained
in this section. The common approach to this problem is to The result of this step is a set of relation clusters for each
let domain experts come up with a small number of rela- association rule. User interaction is needed at this point, in
tions and then to model them in an ontology editor. This order to review the results and to provide meaningful labels
requires knowledge of both ontology creation and ontology for the relations. They are not generated automatically at
editors, which tends to be a too high hurdle for domain ex- the moment, but schemes employing parts-of-speech analysis
perts. Instead, we propose to do it based on the content of (e.g. using the verbs) are feasible.
the corpus in question. With the named entities given by
the preceding steps, relation discovery applying statistical The last step of the algorithm is the transformation of the
methods becomes feasible. entities and their relations into an ontology language. The
transformation process is a straight-forward affair for enti-
ties, classes and binary relations, since those can be handled
5.1 Algorithm by corresponding constructs in RDF. The transformation of
Figure 2 shows the workflow of our approach. The first step, n-ary relations is slightly more complex, since it involves
NER, has been covered already. The next step consists of the blank nodes that act as a hub for the attachment of binary
application of an association rule mining algorithm on the relations to the various members of the relation. The result-
annotated corpus that has been segmented on the sentence- ing RDF represents the ontology for the domain corpus.
level. Only those sentences containing at least two entities
are kept. Each sentence is represented by the set of entity In the use-case of our project, we have to deal with a dy-
classes appearing in it. These item sets serve as input for the namic corpus, since the articles from the wiki are fed back
apriori algorithm[1], that generates a set of association rules into the system to be analyzed. This continually updates the
of the form a → b. Each rule carries two parameters, support semantic network and keeps it on par with the wiki. But an
(the amount of observations supporting it), and confidence additional step is required: relation classification. The rela-
(in our case #(a→b)
#a
). Thresholds for these parameters can tion clusters that have been committed in the initialization
be used to influence the result of the algorithm. phase of the system are used for this task. New instances
of sentences are marked up with named entities and are
The association rules can be ranked according to the two then transformed into word vectors which can be classified
parameters. High support promises higher coverage, high against the relation clusters, and subsequently transformed
confidence hints at a tighter correlation between the entity into RDF. Since the provenance of each triple in the ontol-
classes involved. Rules with more than one succedent tend ogy is known, exchanges can be restricted to those triples
to be more specialized, as evidenced by a higher confidence, that are affected.
and thus offer a higher potential information gain and they
tend to be forgotten by the domain experts, when asked to Preliminary evaluation results of the algorithm show F-mea-
come up with possible relations. sures (F1 = 2∗Recall∗P recision
) between 70% and 75% for
Recall+P recision
clusters representing binary as well as n-ary relations. The
The next step is a clustering phase. It takes an association algorithm usually creates more relation clusters than a hu-
rule as input. The sentences of the rule are preprocessed, i.e. man would, since humans tend to generalize the relations
the named entities are replaced with their respective classes. rather than to have a multitude of minuscule distinctions in
This is done to receive generalized patterns of the relations in their relation set. We have performed an evaluation of the
the sentences. Only the part between the outermost named performance of the algorithm against a part of the corpus
entities is taken and transformed into word vectors. These relevant for the pilot application in the WIKINGER project.
weights of the vectors are created using tf*idf. More details can be found in [2].
The goal of the clustering phase is to receive relation clus-
ters, i.e. clusters in which every vector symbolizes the same 5.2 User interface
In order to provide the domain experts with an interface linguistics. In this respect, WALU complements the range
that facilitates directing the relation discovery process, the of existing tools.
Wikinger Relation Discovery GUI, short WiReD, has been
developed. It allows to view the results of the different steps
of the algorithms and to experiment with different settings
6.2 Ontology learning environments
As has been pointed out above, ontology learning environ-
for them. This encompasses the association rules generated
ments usually are built as supporting tools for ontology
by the apriori algorithm as well as the composition of the
engineers. Their task differs from the one tackled by the
relation clusters generated by the clustering phase.
approaches in this paper insofar as the ontology engineer
has the process-knowledge necessary for building ontologies.
Association rules can be selected manually for clustering,
He usually has access to different domain experts, and thus
clusters can be post-processed (merged with others, deleted,
needs only marginal software support. Named entity recog-
renamed) and finally selected for inclusion into the seman-
nition is employed sometimes to facilitate populating the
tic network. The parameters for each algorithmic step are
ontology, whereas relation discovery is not used extensively,
preset with reasonable defaults, but can be changed directly
at least not to our knowledge.
from within WiReD, thus allowing experiments on the data
set. This may sound intimidating at first reading, but in
Text-To-Onto[11] contains a module that calculates associ-
practice there are never more than two parameters per step
ation rules to provide the engineer with an overview over
in the processing chain, four parameters in total.
possible interrelations between concept classes, but this ap-
proach is not followed further in the context of the applica-
When the experts have come to a final result, i.e. they have
tion. Its successor, Text-2-Onto[5], employs a limited ver-
agreed upon a set of relations they want to see included
sion of relation extraction, insofar as it searches for hyponym
in the ontology, the relation information is fed back into
relation patterns (e.g. ”x is a kind of y”) in order to find ad-
the WIKINGER framework. Here it is used for different
ditional instances of concept classes in a corpus. Relation
purposes. First of all it can be used to transform the infor-
discovery is not employed there.
mation associated with it - the entities and their relations
- into the ontology format of choice. If the corpus is static,
this concludes the work needed for the ontology. In the case 6.3 Relation Discovery
of dynamic corpora, e.g. wiki systems, the relation infor- Hasegawa et al [8] propose a system with a similar approach
mation approved by the experts is used to automatically than the one presented here. They first perform NER on a
classify new patterns that enter the system. These basically text corpus, and then collect entity pairs from within sen-
follow the same steps of the algorithm, only now in a fully tences. These pairs are grouped by composition, the corre-
automated mode. The experts can change the relation set sponding sentences are transformed into word vectors and
anytime they want using the WiReD GUI which results in a a clustering step is performed on each of the groups. This
total recalculation of the ontology to reflect their desire for results in a couple of relation clusters for each group. With
change. some postprocessing (weeding out clusters below a certain
size), they report F-measures of between 75% and 80% for
selected clusters on a year of newspaper articles from The
6. RELATED WORK New York Times. In addition, they generate cluster labels
This section highlights related work in the areas touched by by taking the words with the highest occurrence in each
the work described in the sections above. We concentrate cluster. We believe that adding an association rule creation
on annotation tools rather than individual NER algorithms, phase at the beginning helps in the selection of interesting
since the tools mentioned all encompass different approaches combinations of relation candidates, even more so because
to NER. Following that, ontology learning environments are we are not restricted to the detection of binary relations.
discussed, with a special regard to their use of relation dis-
covery. Finally, algorithms partial to the discipline of rela- There are other approaches besides this one, that exploit
tion discovery are discussed. syntactic structures and perform parts-of-speech analysis:
Jiang et al. [10] analyze sentence grammar trees, model
candidate relations in RDF in order to capture their direc-
6.1 Annotation tools tion and extract from the RDF a set of generalized relations.
As explained in section 4, the rationale behind WALU is
Navigli et al. [14] present an approach to ontology learning
its usability by professionals of any domain, in particular
that exploits synsets from WordNet in order to disambiguate
without computational or linguistic expertise. In this re-
meaning and find relations that might hold between different
spect, WALU differs from other existing tools for semantic
entities from the sentences that explain the different synsets.
annotation, e.g. GATE [7], WordFreak [12], MMAX [13], or
But these approaches are dependent on deeper knowledge of
PALinkA [15]. These tools are primarily intended for users
the language of the text corpus. Approaches like Hasegawa’s
with a background in (computational) linguistics. Conse-
or ours only rely on statistics and the existence of annotated
quently, they are either tailored to different, more complex
entities, thus they are language agnostic.
tasks than WALU (e.g. PALinkA for discourse annota-
tion), or are designed as highly multifunctional tools (e.g.
GATE, WordFreak, or MMAX). This multifunctionality al- 7. FUTURE WORK
lows their flexible application with regard to specific and Regarding NER, we will implement an interface to the Weka
complex needs. However, the price of this flexibility is that library [17], which comprises a number of machine learning
these tools require extensive configuration efforts which sig- algorithms. We will investigate combinations of different
nificantly affects usability for non-experts in computational ML approaches either sequentially (i.e. the output of one
classifier is used as input to another one) or concurrently (i.e. [1] R. Agrawal and R. Srikant. Fast algorithms for mining
several kinds of classifiers are run in parallel and a more-or- association rules. In Proceedings of the 20th VLDB
less sophisticated voting mechanism — which might involve conference, pages 487–499, 1994.
a further ML approach — decides on the final classification). [2] L. Bröcker. Semiautomatic Creation of Semantic
Networks. In Online-proceedings of PhD-symposium at
Furthermore, we plan to provide an interface to the UIMA ESWC 2007, June 2007. no URL as of yet.
framework4 . This way, further facilities for learning and pre- [3] L. Bröcker, M. Rössler, A. Wagner, et al. WIKINGER
processing (e.g. morphological or syntactic analysis, which - Wiki Next Generation Enhanced Repositories. In
can provide useful information for semantic annotation as Online Proceedings of the German E-Science
well as relation discovery) will become available to our frame- Conference, 2007.
work. Since units from the UIMA framework can be pro- [4] N. A. Chinchor, editor. Proceedings of the Seventh
vided as web services they can be added to complement the Message Understanding Conference, Fairfax, VA, 1998.
WIKINGER framework as needed. [5] P. Cimiano and J. Völker. Text-2-Onto. In Proceedings
of NLDB 2005, pages 227–238, 2005.
Regarding relation discovery, we intend to apply our ap-
[6] N. Collier, P. Ruch, and A. Nazarenko, editors.
proach to other data sets, especially from the newspaper
Proceedings of the International Joint Workshop on
domain, in order to evaluate its performance on data sets
Natural Language Processing in Biomedicine and its
that cover a wide range of topics, and to enhance the al-
Applications (JNLPBA-2004), Geneva, Switzerland,
gorithm with a stage that extracts suitable labels for the
2004.
relations and their members automatically.
[7] H. Cunningham. GATE, a General Architectur for
Text Engineering. Computers and the Humanities,
The WIKINGER framework will be developed further, we
36:223–254, 2002.
intend to use it as a base platform for a variety of future
projects. [8] T. Hasegawa, S. Sekine, and R. Grishman. Discovering
Relations among Named Entities from Large Corpora.
In Proceedings of the Annual Meeting of Association of
8. CONCLUSIONS Computational Linguistics, pages 415–422, 2004.
This paper described a new approach to semi-automatic [9] L. Hirschman, A. Yeh, C. Blaschke, and A. Valencia.
knowledge capturing from large text corpora. The goal is to Overview of BioCreAtIvE: critical assessment of
empower domain experts to create domain ontologies them- information extraction for biology. BMC
selves, without being dependent on the availability of on- Bioinformatics, 6 (Supplement 1), 2005.
tology engineers. This is to be achieved by automating the
[10] T. Jiang, A. Tan, and K. Wang. Mining Generalized
process to a high degree, by employing named entity recog-
Associations of Semantic Relations from Textual Web
nition (NER) and relation discovery. Domain experts are in-
Content. IEEE Transactions on Knowledge and Data
volved at those stages which require a substantial knowledge
Engineering, 1(2):164–179, 2007.
of the domain in question. Two software tools aiding in the
process have been introduced that aid the domain experts [11] A. Maedche. The Text-To-Onto Environment, chapter
in the task, WALU and WiReD. The former is a workbench 7 in Alexander Maedche: Ontology Learning for the
for example-based NER, while the latter is a tool aiding in Semantic Web. Kluwer Academic Publishers, 2002.
the relation discovery process. [12] T. Morton and J. LaCivita. WordFreak: an open tool
for linguistic annotation. In Proceedings of the 2003
Evaluation results for the different algorithmic solutions have Conference of the North American Chapter of the
been presented that show high values for F-measure for the Association for Computational Linguistics on Human
automatic knowledge capturing methods. Language Technology, Edmonton, Canada, 2003.
[13] C. Müller and M. Strube. MMAX: A tool for the
All of this is part of a web service based architecture, the annotation of multi-modal corpora. In Proceedings of
WIKINGER framework. It is used to create semantically en- the 2nd IJCAI Workshop on Knowledge and Reasoning
hanced collaborative knowledge platforms for scientific com- in Practical Dialogue Systems, Seattle, WA, 2001.
munities. The pilot application is a semantic wiki for the [14] R. Navigli, P. Velardi, and A. Gangemi. Ontology
domain of contemporary history research regarding German learning and its application to automated terminology
catholicism. translation. IEEE Intelligent Systems, 18(1):22–31,
2003.
9. ACKNOWLEDGMENTS [15] C. Orasan. PALinkA: A highly customisable tool for
The work presented in this paper is being funded by the discourse annotation. In Proceedings of the Fourth
German Federal Ministry of Education and Research under SIGdial Workshop on Discourse and Dialogue,
research grant 01C5965. See http://wikinger-escience.de for Sapporo, Japan, 2003.
further details regarding the project. The authors would [16] A. Wagner and M. Rössler. WALU — Eine
like to thank Prof. Cremers from the University of Bonn Annotations- und Lern-Umgebung für semantisches
and Prof. Hoeppner from the University of Duisburg-Essen Tagging. In G. Rehm, A. Witt, and L. Lemnitzer,
for their helpful suggestions. editors, Data Structures for Linguistic Resources and
Applications, pages 263–271. Gunter Narr Verlag,
Tübingen, 2007.
10. REFERENCES
4
http://incubator.apache.org/uima/
[17] I. H. Witten and F. Eibe. Data Mining: Practical machine
learning tools and techniques. Morgan Kaufmann, San Fran-
cisco, 2nd edition, 2005.