<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Insight to Hyponymy Lexical Relation Extraction in the Patent Genre Versus Other Text Genres</article-title>
      </title-group>
      <contrib-group>
        <aff id="aff0">
          <label>0</label>
          <institution>Linda Andersson</institution>
          ,
          <addr-line>Mihai Lupu, João Pallotti, Florina Piroi, Allan Hanbury</addr-line>
          ,
          <institution>Andreas Rauber Vienna University of Technology Information and Software Engineering Group (IFS) Favoritenstrasse 9-11</institution>
          ,
          <addr-line>1040 Vienna</addr-line>
          ,
          <country country="AT">Austria</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2014</year>
      </pub-date>
      <fpage>8</fpage>
      <lpage>10</lpage>
      <abstract>
        <p>Due to the large amount of available patent data, it is no longer feasible for industry actors to manually create their own terminology lists and ontologies. Furthermore, domain specific thesauruses are rarely accessible to the research community. In this paper we present extraction of hyponymy lexical relations conducted on patent text using lexico-syntactic patterns. We explore the lexico-syntactic patterns. Since this kind of extraction involves Natural Language Processing we also compare the extractions made with and without domain adaptation of the extraction pipeline. We also deployed our modified extraction method to other text genres in order to demonstrate the method's portability to other text domains. From our study we conclude that the lexico-syntactic patterns are portable to domain specific text genre such as the patent genre. We observed that general Natural Language Processing tools, when not adapted to the patent genre, reduce the amount of correct hyponymy lexical relation extractions and increase the number of incomplete extractions. This was also observed in other domain specific text genres.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Patent Text Mining</kwd>
        <kwd>Natural Language Processing</kwd>
        <kwd>Ontology</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        When conducting Prior Art Search it is essential to find different
aspects of a patent? Each aspect can be divided into term pairs
consisting of a general term and a specific term. [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]
This task requires both domain knowledge and access to technical
terminology (both explicit and implicit knowledge). However,
previous studies in the patent genre have observed that patent writers
intentionally use entirely different word combinations to re-create a
“concept”, which increases the vocabulary mismatch issue [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]; and
thereby make commercial technical terminology dictionaries such
as EuroTermBank1 and IATE2 less re-usable [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ].
      </p>
      <p>
        In this paper we explore hyponymy relation extraction from the
collection itself using lexico-syntactic patterns defined in [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ]. With
the variation in concept formulations, where paraphrasing of
existing concepts is generally applied, a support tool such as a thesaurus
or an ontology based on automatic extraction of lexical relations
from the patent genre will be an usable search aid. Automatic
ontology population consists of several steps, normalization of data,
tokenization, Part of Speech (PoS) tagging, etc. However, the
problem of using standard Natural Language Processing (NLP) tools is
that the source data and the target data do not have the same feature
distribution, this being a pre-requisite for their correct use [
        <xref ref-type="bibr" rid="ref26">26</xref>
        ].
Too many unseen events will decrease the performance of broad
coverage NLP tools. In order to reduce the gap between source
and target data several studies involving patent domain adaptation
of broad coverage NLP tools have been investigated [
        <xref ref-type="bibr" rid="ref10 ref11 ref16 ref2 ref23 ref8">16, 10, 2, 11,
23, 8</xref>
        ]. The focus of these adaptations have been either on
reducing the sentence length or increasing the lexicon. Only [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] and [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]
have target adaptations incorporating domain information about the
noun phrases’ (NP) syntactic distributions. In this paper, we re-use
the heuristic rules presented in [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ].
      </p>
      <p>
        The objectives of this study are:
1. to examine if it is possible to extract hyponymy lexical
relations using the general lexico-syntactic patterns defined in
[
        <xref ref-type="bibr" rid="ref5">5</xref>
        ];
2. to verify if the heuristic domain adaptation rules deployed in
the extraction pipeline improve the candidate extractions;
3. to examine the portability of our modified extractor method,
developed for the patent text genre, to other domain specific
genres;
4. to examine if it is possible to simplify the evaluation process
of hyponymy relation extraction.
      </p>
      <p>The remainder of this paper is organized as follows. We first present
some related work and terminology in Section 2. In Section 3 we
present our experimental set up. In Section 4 we report our general
results. Section 5 presents our conclusion and future work.
1http://project.eurotermbank.com/
2http://iate.europa.eu/</p>
    </sec>
    <sec id="sec-2">
      <title>2. RELATED WORK</title>
      <p>In the Information Retrieval (IR) community, the patent retrieval
research has focused mainly on improvements and method
developments within systems for supporting patent experts in the process
of Prior Art search. Less research attention has been given to other
type of resources that support the patent examiner in the
information process activities.</p>
    </sec>
    <sec id="sec-3">
      <title>2.1 Terminology Effect on NLP</title>
      <p>
        Before we can re-use NLP tools in text genres with high density of
scientific terminology and new words, we need to understand the
word formation process of the English language. The most
productive word formation in English is affixation i.e. adding prefixes or
suffixes to a root [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. The suffixes ‘-ing’ and ‘-ed’ are especially
problematic for NLP applications because when they are added to
verbs, the new formed word may be a noun, an adjective or remain
a verb (as in sentence 8, Figure 1 in the Appendix).
      </p>
      <p>
        One of the major mechanisms of word formation is the
morphological composite, which allows the formation of compound nouns out
of two nouns (e.g. floppy disk, air flow) [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ], and thereby creating
a Multi Word Unit (MWU). It has been observed that in the
technical jargon a heavy use of noun compounds constitutes the
majority of scientific terminologies [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ]. The compounding strategy
causes not only unseen events on word level with new
orthographical units, it also generates a diversity of syntactic structures among
noun phrases, which is problematic for NLP tools [
        <xref ref-type="bibr" rid="ref10 ref24">10, 24</xref>
        ].
Furthermore, many NLP applications have chosen to overlook MWUs
due to their complexity and flexible nature [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ].
      </p>
      <p>
        NPs can consist of single tokens, or can as well be as long and
complex as any other occurring phrases in a sentence [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ]. The NPs
have an internal structure that dictates where additional elements
can occur in relation to the head noun (e.g. pre- and post-modifier).
There is a range of elements that can take the pre-modifier role in an
NP but adjectives are the most typical pre-modifiers. In hyponymy
lexical relation extraction, adjectives have a semantic significance,
since the adjective modifiers could be considered a hyponym to
the head noun [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. For example, ‘apple juice’ is a valid hyponym
to ‘juice’, but only in this combination since the modifier ‘apple’
specifies the head ‘juice’ [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. The post-modifier construction is
more complex, since a head noun can be post-modified by both
phrases and clauses.
      </p>
      <p>
        One central concept when analyzing NPs is to define its head [
        <xref ref-type="bibr" rid="ref24">24</xref>
        ].
The head in an NP has a supreme importance, as is the central part
of the noun (e.g.“the poet Burns”, “Burns the poet”) [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ]. When a
NP contains a prepositional phrase the traditional linguists promote
the proper name (e.g. “the city of Stockholm”) or the NP followed
after the preposition (e.g. “a group of DNA strings”) as the main
head noun, since the NP after the preposition tends to have the
highest degree of lexicalization [
        <xref ref-type="bibr" rid="ref15 ref24 ref6">6, 24, 15</xref>
        ]. However, what should be
identified as the head noun in an NP is not straight forward [
        <xref ref-type="bibr" rid="ref24">24</xref>
        ].
Moreover, in [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] it was observed that the syntactic parsers
rightheaded bias caused problems during the analysis step of the patent
sentences, thereby yielding erroneous analyzes.
      </p>
    </sec>
    <sec id="sec-4">
      <title>2.2 Patent Text Effects in NLP</title>
      <p>
        Patents are semi-structured documents which offer many
different applications for text mining [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. In patent documents, abstract
and non-standard terminology is used to avoid narrowing the scope
of the invention, unlike the style of other genres like newspapers
and scientific articles [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ]. Moreover, the vocabulary varies over
time with terms such as “LP” and “water closet” being regarded
as instances of obsolescence [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. This type of discourse
characteristic makes the patent text mining task more challenging. Many
Patent Retrieval studies have tried to address different patent search
problems by applying linguistic knowledge, using broad coverage
NLP tools. However, as the generic NLP tools are not trained on
the patent domain they experience problems with parsing long and
complex NPs [
        <xref ref-type="bibr" rid="ref10 ref8">10, 8</xref>
        ]. There have been several studies focusing on
reducing the gap between the source and target data, the focus
being placed mainly on sentence reduction [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ], on lexicon increase
[
        <xref ref-type="bibr" rid="ref16">16</xref>
        ], or on both [
        <xref ref-type="bibr" rid="ref23">23</xref>
        ]. However, just increasing the lexical coverage
or decomposing sentences will not solve the problem, since token
coverage and sentence length are only part of the problem. [
        <xref ref-type="bibr" rid="ref28">28</xref>
        ]
concluded that, since there is no significant difference between the
general English and the English used in the patent discourse, on
single token coverage, the technical terminology is more likely present
in multi-word constructions consisting of complex NPs.
Information about NPs’ syntactic distribution has only been deployed in [
        <xref ref-type="bibr" rid="ref2 ref8">2,
8</xref>
        ], in order to improve the NLP analysis. In [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] a hierarchical
chunker was designed to fit the syntactic structure of the patent sentence,
targeting embedded NPs, while in [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] heuristic rules addressing the
most common observed errors made by the NLP tools were used as
a post correcting filter.
      </p>
    </sec>
    <sec id="sec-5">
      <title>2.3 Ontology Population</title>
      <p>
        Automatic ontology population relates to the methods used in
Information Extraction (IE) as the general purpose is to extract
predefined relations from text, hence referred to ontology based
information extraction (OBIE) [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ]. There are several applications
where OBIE is used to enhance domain knowledge, to create a
customized ontology, and in rich existing ontologies. OBIE
techniques consist of identifying named entities (NE), technical terms,
or relations. The OBIE process consists of several steps, data
normalization, tokenization, PoS tagging, etc., thereafter following the
recognition steps like gazetteers combined with rule-based
grammars, ontology design pattern (ODP), pattern slots identifications
such as lexico-syntactic pattern (LSP). Different techniques for
hyponymy lexical relation extraction have been explored âA˘ S¸ many
of them depending on pre-encoded knowledge such as domain
ontologies and machine readable dictionaries [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. In order to avoid the
need of pre-existing domain knowledge and remain independent of
the sub-language one option is to use generic LSPs for hyponymy
lexical relation extraction. [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] proposed a method to extract
hyponymy lexical relations based on five LSPs, see Table 1.
There are several issues related to extracting relations from a raw
text based on LSPs. For instance, the LSP examples 2, 5 and 6
in Table 1 are not clear cases of hyponymy lexical relations, as in
‘domestic pets such as cats and dogs,’ since in LSP 2 Germany,
France and Italy are members of the European Community and in
LSP 6 France, England and Spain are countries in Europe i.e. a part
of the geographic content called Europe [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ].
      </p>
      <p>
        With a wider semantic definition of the hyponym property, we can
include both ‘part of’ and ‘member of’ in the definition:
“: : : an expression A is a hyponym of an expression B iff the
meaning of B is part of the meaning of A and A is subordinated of B. In
addition to the meaning of B, the meaning of A must contain
further specifications, rendering the meaning of A, the hyponym, more
specific than the meaning of B. If A is a hyponym of B, B is called
a hypernym of A.” [18, p83]
Hearst’s patterns, [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ], give high precision but low recall, while
ODP gives high recall and low precision [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ]. In [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ], LSP 1
was used to extract candidate relations from the Grolier’s
American Academic Encyclopaedia (8.6M words). In this study, 7,067
sentences match LSP 1 and 152 relations fit the restriction i.e. to
contain an unmodified noun (or with just one modifier).
A common approach to evaluate hyponymy relation extractions is
to use an existing ontology as a gold standard [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. For instance, in
[
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] the assessment was conducted by looking up if the relation was
found in WordNet. Out of 226 unique words, 180 words existed in
the WordNet hierarchy, and 61 out of 106 relations already existed
in the WordNet. However, since most of the terms in WordNet are
unmodified nouns or nouns with a single modifier, using WordNet
in the evaluation process of this study was not feasible.
In [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] the gold standard was created by using linguists, but this type
of labeling task is both time-consuming and costly, which makes
the approach feasible only for small gold standards. The
annotators were asked to manually identify domain-specific terms, NEs,
synonymy and hyponymy relationships between identified terms
and NEs. The annotation task requires both linguistic knowledge,
as well as, some domain specific knowledge.
      </p>
      <p>
        The gold standard was used to evaluate automatic hyponymy
relation extractions from technical corpora, in English and Dutch. The
data consisted of dredging year reports and news articles from the
financial domain. The data was enriched with PoS tagging and
lemmas produced by the LeTs Preprocessing Toolkit. The LeTs
Preprocessing toolkit was trained on similar data where the
accuracy of the PoS tagger was 96.3%. The NE extractor only achieved
a recall of 62.92% and a precision of 59.33% [
        <xref ref-type="bibr" rid="ref27">27</xref>
        ].
      </p>
      <p>
        For the hyponymy lexical relations extraction, three different
techniques were used: 1) a lexico-syntactic pattern model based on LSP
in [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ], 2) a distribution model using context cluster by an
agglomerative clustering technique and 3) a morpho-syntactic model. The
morpho-syntactic model is based on the head-modifier principle:
Single-word NP, if lexical item L0 is a suffice string of lexical
item L1, L0 is a hypernym of L1
MWUs NP, if lexical item L0 is the head of term of lexical
item L1, then L0 is a hypernym of L1
NP + prepositional phrase, if lexical item L0 is the first part
of a term in L1 containing a NP plus prepositions (EN: of,
for, before, from, to, on), then L0 is to be the hypernym of
L1.
[
        <xref ref-type="bibr" rid="ref17">17</xref>
        ] concluded that the pattern-based methods and especially the
morpho-syntactic approach achieved good performance on the
technical domain data, therefore demonstrating that the general purpose
hypernym detection models are portable to other domain and
userspecific data.
      </p>
      <p>
        In [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ], hyponymy relations were extracted from US and Japanese
patent re-using LSP patterns in [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ]. For English 3,898,000 and
for Japanese 7,031,149 candidate hyponymy relations were
identified. The alignment between the language pair was conducted via
citation analysis; 2,635 pairs of English-Japanese hyponymy
relations were manually evaluated. The best method obtained Recall
of 79.4% and Precision of 77.5%.
      </p>
    </sec>
    <sec id="sec-6">
      <title>3. OUR APPROACH</title>
      <p>Our data sets consist of five different text genres: the Brown
corpus3 (henceforth Brown), the WO and EP patent documents of
IREC (Patent)4, the TREC test collection for Clinical Decision
Support Track (MedIR)5, the test collection for Mathematical retrieval
provided by NTCIR (MathIR)6, and the papers produced during the
Conference and Labs of Evaluation forum7 (CLEFpaper). In Table
2 we present the total amount of sentences fitting the LSPs per data
and extraction methods.</p>
    </sec>
    <sec id="sec-7">
      <title>3.1 Method</title>
      <p>
        For this experiment we applied exactly the same methodology to all
5 data sets. We used all of the LSP patterns in Table 1. For the NLP
pipeline we enriched all data sets with PoS tags using the Stanford
tagger – English-left3words-distisim.tagger model [
        <xref ref-type="bibr" rid="ref25">25</xref>
        ]. In order to
allow more flexibility to the phrase boundary we chose to use the
baseNP Chunker [
        <xref ref-type="bibr" rid="ref22">22</xref>
        ]. We defined three pipeline extraction
methods:
1. No rules (NoRules) was used to modifying the NLP pipeline
analyzes
2. Three rules (SimpleRules) addressing observed errors among
sentence fitted the LSP patterns. The rules address different
type of conjunction and commas issues. Rule i) NP [cat and
dogs] changed to two NPs [cat] and [dog], ii) [cat or dogs]
changed two NPs [cat] or [dog], iii) numerous listing with
commas.
3http://www.hit.uib.no/icame/brown/bcm.html
4IREC, is the corrected version of the MAREC http://www.ifs.tuwien.ac.
at/imp/marec.shtml
5http://www.trec-cds.org/2014.html
6http://ntcir-math.nii.ac.jp/
7http://www.clef-initiative.eu/publication/proceedings
3. Domain rules, (DomainRules) here we applied the simple
rules (2) and the rules presented in [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ].
For the evaluation only a smaller set was sampled out (1,647
instances) for manual assessment, approximately 100 instances per
data collection and method. One instance correspond to one
relation extracted from a sentences, if there are several possible
extraction in a single sentence, each extraction correspond to one
instance (see figure 1 in the appendix). Therefore not exact 1,500
instances were evaluated since some sentences contain more than
one instances. Due to the fact that there are very few people
having the level of linguistic knowledge, as well as the domain
specific knowledge required to conduct assessment, we decided upon
a more generic evaluation schema. The assessors were divided into
three groups: linguist, and expert and non-expert. The linguist has
domain knowledge of the patent domain and the computer science
domain.
      </p>
      <p>
        For the evaluation task, we constructed a simple interface, see
figure 1 in the appendix. The evaluation tool shows the original
sentence and five definition of relations between L0 and L1; i) L0 is
a kind of L1, ii) L0 is a part of L1, iii) L0 is a member of L1; iv)
L0 is in another relation with L1, v) L0 has no relation to L1. For
uncertainty of the assessor we added Cannot say anything about the
two and for erroneous extraction we added The sentence makes no
sense. Since the NP boundaries were not entirely correct identified
for all extractions, we added a check box for wrong boundary (for
L0 and L1). In the instruction for the evaluation task, a simple
example and a domain example were given for all types of relations.
In order to find out how difficult the task was thought to be by the
assessors, we asked each assessor to grade each relation from as
scale 1 (very easy) to 5 (very difficult). Furthermore, since it was
observed in [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] that web searches for many candidate phrases were
required in order to understand their meaning, we gave the assessor
the possibility to search for the concept via a web service. We aim
to improve the evaluation tool and give better interactive support
therefore this feedback information is valuable for us.
      </p>
    </sec>
    <sec id="sec-8">
      <title>4. RESULTS</title>
      <p>In Table 3 we present the evaluation result based upon the linguist
assessor. We see that the NoRules method generates more
candidate extractions compared to the other ones, with correct boundary
identification. This fact puzzled us since our experience during the
assessment indicated the opposite. For instance, a common error
was deverbal nouns exclusion. This error especially decreased
correct and complete extractions for the domain specific text genres
when using NoRules. For instance, when the head noun is a
deverbal noun, the PoS-tagger assigns the label verb instead of a noun
(e.g. “ultrasonic/JJ welding/VBN” and “laser/NN welding/VBN”,
and compare sentences 7 and 8 in figure 1, appendix).</p>
      <p>Our first assumption to this contradiction was that one of the rules
in the DomainRule method, which unifies NPs with ‘of’-construction,
harmed the extractions. In example 1, the hypernym consists of an
embedded NP with prepositional ‘of’-construction modifying the
head noun.</p>
      <p>Example 1: Embedded NP ‘of’-construction
The novel conjugate molecules are provided for the manufacture of a medicament
for gene therapy, apoptosis, or for the treatment of diseases such as cancer,
autoimmune diseases or infectious diseases.</p>
      <p>If we include the entire NP i.e. “the treatment of diseases” the
hyponymy lexical relation becomes incorrect since “cancer”,
“autoimmune diseases” and “infectious diseases” are “diseases” and
not “treatments”. On the other hand, in sentence 5 (figure 1,
appendix) the relation between the hypernym and hyponyms becomes
incorrect since hyponyms constitute properties of the hyponym
therefore the NP should be unified. In sentences 3 and 4 (figure 1,
appendix) the unification of the NPs ‘of’-construction is more
doubtful for the hypernym where “potential risk factors” (sentence 3)
compared to “the distribution of potential risk factors” (sentence
4) seems to be the better choice. However, one of the hyponyms is
overlooked in sentence 3 but extracted in sentence 4 with the help of
the domain rule unifying ‘of’-construction NPs. When examining
the outcome of the rule we found that 131 instances were
considered correct (i.e. the NP with ‘of’-construction should be unified)
and only 44 instances were incorrect. The more likely reason for
the NoRules more complete and correct identified hyponymy
relation is that the NoRules generated more extractions compared to
DomainRules which has a more strict extraction rule schema.
Table 4 shows the percentage of the most dominant relation “a Kind
Of” and all positive relations (“a Kind Of”, “a Part Of”, “a
Member Of”, “another Relation”) for each method and data set. The
preferred hypernym rule is the DomainRules method regardless of
data set. For hyponyms, the result is more inconclusive since
several methods ended up having the same percentages. For the “a
Kind Of” relation the preferred method is either SimpleRules or
NoRules as seen in Table 4.</p>
      <p>Table 5 displays the percentage of all examined sentences matching
the LSP patterns where a positive and correct extraction was
identified. For three out of five data sets the method SimpleRules was</p>
      <p>In order to examine the simplification of the evaluation process, we
computed inter-annotation agreements between the three groups:
expert, linguist and non-expert. The inter-annotation agreement
for identifying relations ranges between 81% and 88% (Table 6),
regardless of the group comparisons for Brown and for the
scientific paper data sets. Similar agreement values were found for the
patent and medical text domain. The inter-annotation agreement
decreases for wrong NP boundary identifications, which can be
explained by that fact that it requires linguistic schooling to correctly
identify NPs.</p>
    </sec>
    <sec id="sec-9">
      <title>5. CONCLUSIONS</title>
      <p>We conclude the following:</p>
      <p>
        It is possible to re-use LSPs for hyponymy lexical relation
extractions. We thereby confirm the observation made in [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]
that the LSP method for relation extraction is portable to
different text genres.
      </p>
      <p>We also confirm that for domain specific text genre, such as
patent or medical genres, at least for the hypernyms
modification of NLP tools is required. For detecting hyponyms
the additional rules were less successful. On the other hand,
as seen in sentences 7 and 8 (figure 1, appendix) the rules
addressing deverbal nouns make it possible to extract more
correct instances.</p>
      <p>The simplified process of evaluating hyponymy lexical
relations extractions using non-linguists and non-experts is on
an acceptable inter-annotation agreement level. However,
more information regarding the identification of NP
boundaries should be added in future evaluation guidelines.</p>
      <p>
        In the future we will explore machine learning algorithms
to select which extraction method should be used for a
specific relation, instance and data collection. The additional
modifying the NLP pipeline need further examination, since
it becomes contra productive for some instance but improve
for others. Furthermore, we also want to examine additional
patterns exploring similarity between the internal structures
of NPs, as described in [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ].
      </p>
      <p>
        In the future we will explore machine learning algorithms to
select which extraction method should be used for a specific relation,
instance and data collection. The additional modifying the NLP
pipeline need further examination, since it becomes contra
productive for some instance but improve for others. Furthermore, we also
want to examine additional patterns exploring similarity between
the internal structures of NPs, as described in [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ].
      </p>
    </sec>
    <sec id="sec-10">
      <title>6. ACKNOWLEDGMENTS</title>
      <p>This research was partly funded by the Austrian Science Fund (FWF)
projects P25905-N23 (ADmIRE) and I1094-N23 (MUCKE).</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>A.</given-names>
            <surname>Adams</surname>
          </string-name>
          .
          <article-title>Personal correspondence</article-title>
          .
          <source>PatOlympics</source>
          ,
          <year>2011</year>
          , Vienna,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>L.</given-names>
            <surname>Andersson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Lupu</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <given-names>A.</given-names>
            <surname>Hanbury</surname>
          </string-name>
          .
          <article-title>Domain adaptation of general natural language processing tools for a patent claim visualization system</article-title>
          . In M. Lupu, E. Kanoulas, and F. Loizides, editors,
          <source>Multidisciplinary Information Retrieval</source>
          , volume
          <volume>8201</volume>
          of Lecture Notes in Computer Science, pages
          <fpage>70</fpage>
          -
          <lpage>82</lpage>
          . Springer Berlin Heidelberg,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>P.</given-names>
            <surname>Anick</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Verhagen</surname>
          </string-name>
          , and
          <string-name>
            <given-names>J.</given-names>
            <surname>Pustejovsky</surname>
          </string-name>
          .
          <article-title>Identification of multiword expressions in the brwac</article-title>
          . In N. Calzolari,
          <string-name>
            <given-names>K.</given-names>
            <surname>Choukri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Declerck</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Loftsson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Maegaard</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Mariani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Moreno</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Odijk</surname>
          </string-name>
          , and S. Piperidis, editors,
          <source>Proceedings of LREC-2014</source>
          , Reykjavik, Iceland, may
          <year>2014</year>
          .
          <article-title>European Language Resources Association (ELRA).</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>V.</given-names>
            <surname>Arranz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Atserias</surname>
          </string-name>
          , and
          <string-name>
            <given-names>M.</given-names>
            <surname>Castillo</surname>
          </string-name>
          .
          <article-title>Multiwords and word sense disambiguation</article-title>
          . In A. Gelbukh, editor,
          <source>Computational Linguistics and Intelligent Text Processing</source>
          , volume
          <volume>3406</volume>
          of Lecture Notes in Computer Science, pages
          <fpage>250</fpage>
          -
          <lpage>262</lpage>
          . Springer Berlin Heidelberg,
          <year>2005</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>K. H.</given-names>
            <surname>Atkinson</surname>
          </string-name>
          .
          <article-title>Toward a more rational patent search paradigm</article-title>
          .
          <source>In Proceedings of the 1st ACM Workshop on Patent Information Retrieval</source>
          ,
          <source>PaIR '08</source>
          , pages
          <fpage>37</fpage>
          -
          <lpage>40</lpage>
          , New York, NY, USA,
          <year>2008</year>
          . ACM.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>K.</given-names>
            <surname>Ballard</surname>
          </string-name>
          .
          <source>The Frameworks of English. Palgrave Macmillan</source>
          ,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>W.</given-names>
            <surname>Bosma</surname>
          </string-name>
          and
          <string-name>
            <given-names>P.</given-names>
            <surname>Vossen</surname>
          </string-name>
          .
          <article-title>Bootstrapping language neutral term extraction</article-title>
          . In N. Calzolari,
          <string-name>
            <given-names>K.</given-names>
            <surname>Choukri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Maegaard</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Mariani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Odijk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Piperidis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Rosner</surname>
          </string-name>
          , and D. Tapias, editors,
          <source>Proceedings of the Seventh International Conference on Language Resources and Evaluation (LREC'10)</source>
          , Valletta, Malta, may
          <year>2010</year>
          .
          <article-title>European Language Resources Association (ELRA).</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>N.</given-names>
            <surname>Bouayad-Agha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Burga</surname>
          </string-name>
          , G. Casamayor,
          <string-name>
            <given-names>J.</given-names>
            <surname>Codina</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Nazar</surname>
          </string-name>
          , and
          <string-name>
            <given-names>L.</given-names>
            <surname>Wanner</surname>
          </string-name>
          .
          <article-title>An exercise in reuse of resources: Adapting general discourse coreference resolution for detecting lexical chains in patent documentation</article-title>
          . In N. Calzolari,
          <string-name>
            <given-names>K.</given-names>
            <surname>Choukri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Declerck</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Loftsson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Maegaard</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Mariani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Moreno</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Odijk</surname>
          </string-name>
          , and S. Piperidis, editors,
          <source>Proceedings of LREC-2014)</source>
          , Reykjavik, Iceland, may
          <year>2014</year>
          .
          <article-title>European Language Resources Association (ELRA).</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>P.</given-names>
            <surname>Cimiano</surname>
          </string-name>
          .
          <source>Ontology Learning and Population from Text: Algorithms, Evaluation and Applications</source>
          . Springer-Verlag New York, Inc., Secaucus, NJ, USA,
          <year>2006</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>E. D'hondt</surname>
            , S. Verberne,
            <given-names>C. H. A.</given-names>
          </string-name>
          <string-name>
            <surname>Koster</surname>
            , and
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Boves</surname>
          </string-name>
          .
          <article-title>Text representations for patent classification</article-title>
          .
          <source>Computational Linguistics</source>
          ,
          <volume>39</volume>
          (
          <issue>3</issue>
          ):
          <fpage>755</fpage>
          -
          <lpage>775</lpage>
          ,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>G.</given-names>
            <surname>Ferraro</surname>
          </string-name>
          .
          <article-title>Towards deep content extraction from specialized discourse: the case of verbal relations in patent claims</article-title>
          .
          <source>PhD thesis</source>
          , Universitat Pompeu Fabra,
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>C. G.</given-names>
            <surname>Harris</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Arens</surname>
          </string-name>
          , and
          <string-name>
            <given-names>P.</given-names>
            <surname>Srinivasan</surname>
          </string-name>
          .
          <article-title>Using classification code hierarchies for patent prior art searches</article-title>
          . In M. Lupu,
          <string-name>
            <given-names>K.</given-names>
            <surname>Mayer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Tait</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <surname>A. J</surname>
          </string-name>
          . Trippe, editors,
          <source>Current Challenges in Patent Information Retrieval</source>
          , volume
          <volume>29</volume>
          <source>of The Information Retrieval Series</source>
          , pages
          <fpage>287</fpage>
          -
          <lpage>304</lpage>
          . Springer Berlin Heidelberg,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>M. A.</given-names>
            <surname>Hearst</surname>
          </string-name>
          .
          <article-title>Automatic acquisition of hyponyms from large text corpora</article-title>
          .
          <source>In Proceedings of the 14th Conference on Computational Linguistics - Volume 2, COLING '92</source>
          , pages
          <fpage>539</fpage>
          -
          <lpage>545</lpage>
          , Stroudsburg, PA, USA,
          <year>1992</year>
          .
          <article-title>Association for Computational Linguistics</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>J. S.</given-names>
            <surname>Justeson</surname>
          </string-name>
          and
          <string-name>
            <given-names>S. M.</given-names>
            <surname>Katz</surname>
          </string-name>
          .
          <article-title>Technical terminology: some linguistic properties and an algorithm for identification in text</article-title>
          .
          <source>Natural Language Engineering</source>
          ,
          <volume>1</volume>
          :
          <fpage>9</fpage>
          -
          <issue>27</issue>
          , 3
          <year>1995</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <surname>E. Keizer.</surname>
          </string-name>
          <article-title>The English Noun Phrase: The nature of linguistic categorization</article-title>
          . Cambridge University Press,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>C. H.</given-names>
            <surname>Koster</surname>
          </string-name>
          and
          <string-name>
            <given-names>J. G.</given-names>
            <surname>Beney</surname>
          </string-name>
          .
          <article-title>Phrase-based document categorization revisited</article-title>
          .
          <source>In Proceedings of the 2Nd International Workshop on Patent Information Retrieval</source>
          ,
          <source>PaIR '09</source>
          , pages
          <fpage>49</fpage>
          -
          <lpage>56</lpage>
          , New York, NY, USA,
          <year>2009</year>
          . ACM.
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>E.</given-names>
            <surname>Lefever</surname>
          </string-name>
          , M. V. de Kauter, and
          <string-name>
            <given-names>V.</given-names>
            <surname>Hoste</surname>
          </string-name>
          .
          <article-title>Evaluation of automatic hypernym extraction from technical corpora in english and dutch</article-title>
          . In N. Calzolari,
          <string-name>
            <given-names>K.</given-names>
            <surname>Choukri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Declerck</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Loftsson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Maegaard</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Mariani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Moreno</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Odijk</surname>
          </string-name>
          , and S. Piperidis, editors,
          <source>Proceedings of LREC-2014</source>
          , pages
          <fpage>490</fpage>
          -
          <lpage>497</lpage>
          ,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>S. Löbner. Understanding</given-names>
            <surname>Semantics</surname>
          </string-name>
          . Oxford University Press, New York,
          <year>2002</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>D.</given-names>
            <surname>Maynard</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Li</surname>
          </string-name>
          , and
          <string-name>
            <given-names>W.</given-names>
            <surname>Peters</surname>
          </string-name>
          .
          <article-title>Nlp techniques for term extraction and ontology population</article-title>
          .
          <source>In Proceedings of the 2008 Conference on Ontology Learning and Population: Bridging the Gap Between Text and Knowledge</source>
          , pages
          <fpage>107</fpage>
          -
          <lpage>127</lpage>
          , Amsterdam, The Netherlands, The Netherlands,
          <year>2008</year>
          . IOS Press.
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>V.</given-names>
            <surname>Mititelu</surname>
          </string-name>
          .
          <article-title>Automatic Extraction of Patterns Displaying Hyponym-Hypernym Co-Occurrence from Corpora</article-title>
          .
          <source>In Proceedings of the 1st CESCL</source>
          , Budapest, Hungary,
          <year>2006</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>H.</given-names>
            <surname>Nanba</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Mayumi</surname>
          </string-name>
          , and
          <string-name>
            <given-names>T.</given-names>
            <surname>Takezawa</surname>
          </string-name>
          .
          <article-title>Automatic construction of a bilingual thesaurus using citation analysis</article-title>
          .
          <source>In Proceedings of the 4th Workshop on Patent Information Retrieval</source>
          ,
          <source>PaIR '11</source>
          , pages
          <fpage>25</fpage>
          -
          <lpage>30</lpage>
          , New York, NY, USA,
          <year>2011</year>
          . ACM.
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>L. A. R.</given-names>
            and
            <surname>M. P. M.</surname>
          </string-name>
          <article-title>Text chunking using transformation-based learning</article-title>
          . In S. Armstrong,
          <string-name>
            <given-names>K.</given-names>
            <surname>Church</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Isabelle</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Manzi</surname>
          </string-name>
          , E. Tzoukermann, and D. Yarowsky, editors,
          <source>Natural Language Processing Using Very Large Corpora</source>
          , volume
          <volume>11</volume>
          of Text,
          <source>Speech and Language Technology</source>
          , pages
          <fpage>157</fpage>
          -
          <lpage>176</lpage>
          . Springer Netherlands,
          <year>1999</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>S.</given-names>
            <surname>Sheremetyeva</surname>
          </string-name>
          .
          <article-title>Natural language analysis of patent claims</article-title>
          .
          <source>In Proceedings of the ACL-2003 Workshop on Patent Corpus Processing -</source>
          Volume
          <volume>20</volume>
          , PATENT '
          <volume>03</volume>
          , pages
          <fpage>66</fpage>
          -
          <lpage>73</lpage>
          , Stroudsburg, PA, USA,
          <year>2003</year>
          .
          <article-title>Association for Computational Linguistics</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>K. Spärck</given-names>
            <surname>Jones</surname>
          </string-name>
          .
          <article-title>Compound Noun Interpretation Problems</article-title>
          . In F. Fallside and W. A. Woods, editors,
          <source>Computer Speech Processing</source>
          . Prentice-Hall, Englewood Cliffs, NJ,
          <year>1983</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>K.</given-names>
            <surname>Toutanova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Klein</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. D.</given-names>
            <surname>Manning</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Y.</given-names>
            <surname>Singer</surname>
          </string-name>
          .
          <article-title>Feature-rich part-of-speech tagging with a cyclic dependency network</article-title>
          .
          <source>In Proceedings of the 2003 Conference of the North American Chapter of the Association for Computational Linguistics on Human Language Technology - Volume 1, NAACL '03</source>
          , pages
          <fpage>173</fpage>
          -
          <lpage>180</lpage>
          , Stroudsburg, PA, USA,
          <year>2003</year>
          .
          <article-title>Association for Computational Linguistics</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [26]
          <string-name>
            <given-names>J.</given-names>
            <surname>Turmo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Ageno</surname>
          </string-name>
          , and
          <string-name>
            <given-names>N.</given-names>
            <surname>Català</surname>
          </string-name>
          .
          <article-title>Adaptive information extraction</article-title>
          .
          <source>ACM Comput. Surv.</source>
          ,
          <volume>38</volume>
          (
          <issue>2</issue>
          ),
          <year>July 2006</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [27]
          <string-name>
            <surname>M. van de Kauter</surname>
            , G. Coorman,
            <given-names>E.</given-names>
          </string-name>
          <string-name>
            <surname>Lefever</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Desmet</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Macken</surname>
            , and
            <given-names>V.</given-names>
          </string-name>
          <string-name>
            <surname>Hoste</surname>
          </string-name>
          . LeTs Preprocess:
          <article-title>The multilingual LT3 linguistic preprocessing toolkit</article-title>
          .
          <source>Computational Linguistics in the Netherlands Journal</source>
          ,
          <volume>3</volume>
          :
          <fpage>103</fpage>
          -
          <lpage>120</lpage>
          ,
          <issue>12</issue>
          /
          <year>2013</year>
          2013.
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          [28]
          <string-name>
            <given-names>S.</given-names>
            <surname>Verberne</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. H. A.</given-names>
            <surname>Koster</surname>
          </string-name>
          , and
          <string-name>
            <given-names>N.</given-names>
            <surname>Oostdijk</surname>
          </string-name>
          .
          <article-title>Quantifying the challenges in parsing patent claims</article-title>
          .
          <source>In In Proceedings of the 1st International Workshop on Advances in Patent Information Retrieval (AsPIRe</source>
          <year>2010</year>
          , pages
          <fpage>14</fpage>
          -
          <lpage>21</lpage>
          ,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>