=Paper= {{Paper |id=None |storemode=property |title=Integrating Open and Closed Information Extraction: Challenges and First Steps |pdfUrl=https://ceur-ws.org/Vol-1064/Dutta_Integrating.pdf |volume=Vol-1064 |dblpUrl=https://dblp.org/rec/conf/semweb/DuttaMNP13 }} ==Integrating Open and Closed Information Extraction: Challenges and First Steps== https://ceur-ws.org/Vol-1064/Dutta_Integrating.pdf
      Integrating Open and Closed Information
       Extraction: Challenges and First Steps

Arnab Dutta1 , Mathias Niepert2 , Christian Meilicke1 , Simone Paolo Ponzetto1
       1
           Research Data and Web Science, University of Mannheim, Germany
       2
           Computer Science and Engineering, University of Washington, USA
              {arnab,christian,simone}@informatik.uni-mannheim.de
                           mniepert@cs.washington.edu



      Abstract. Over the past years, state-of-the-art information extraction
      (IE) systems such as NELL [5] and ReVerb [9] have achieved impres-
      sive results by producing very large knowledge resources at web scale
      with minimal supervision. However, these resources lack the schema in-
      formation, exhibit a high degree of ambiguity, and are difficult even
      for humans to interpret. Working with such resources becomes easier
      if there is a structured information base to which the resources can be
      linked. In this paper, we introduce the integration of open information
      extraction projects with Wikipedia-based IE projects that maintain a
      logical schema, as an important challenge for the NLP, semantic web,
      and machine learning communities. We describe the problem, present a
      gold-standard benchmark, and take the first steps towards a data-driven
      solution to the problem. This is especially promising, since NELL and
      ReVerb typically achieve a very large coverage, but still still lack a full-
      fledged clean ontological structure which, on the other hand, could be
      provided by large-scale ontologies like DBpedia [2] or YAGO [13].

      Keywords: Information extraction, Entity Linking, Ontologies


1   Introduction

Research on information extraction (IE) systems has experienced a strong mo-
mentum in recent years. While Wikipedia-based information extraction projects
such as DBpedia [1, 17] and YAGO [25, 13] have been in development for sev-
eral years, systems such as NELL [5] and ReVerb [9] that work on very large
and unstructured text corpora have more recently achieved impressive results.
The developers of the latter systems have coined the term open information
extraction (OIE), to describe information extraction systems that are not con-
strained by the boundaries of encyclopedic knowledge and the corresponding
fixed schemata that are, for instance, used by YAGO and DBpedia. The data
maintained by OIE systems is important for analyzing, reasoning about, and
discovering novel facts on the web and has the potential to result in a new gen-
eration of web search engines [7]. At the same time, the data of open IE projects
would benefit from a corresponding logical schema even if it was incomplete and
2       A. Dutta et al.

light-weight in nature. Hence, we believe that the problem of integrating open
and schema-driven information extraction projects is a key scientific challenge.
In order to integrate existing IE projects we have to overcome a difficult prob-
lem of linking different manifestations of the same real world object, or more
commonly the task of entity resolution. The fact that makes this task challeng-
ing is that triples from such systems are underspecified and ambiguous. Let us
illustrate this point with an example triple from Nell where two terms (subject
and object) are linked by some relationship (predicate):
                   agentcollaborateswithagent(royals, mlb)
In this triple, royals and mlb are two terms which are linked by some relation
agentcollaborateswithagent. Interpreting these terms is difficult since they
can have several meanings, including very infrequent and highly specialized ones,
which are sometimes difficult to interpret even for humans. Here, royals refers
to the baseball team Kansas City Royals and mlb to Major League Baseball.
    In general, due to the fact that information on the Web is highly hetero-
geneous, there can be a fair amount of ambiguity in the extracted facts. The
problem becomes even more obvious when we encounter triples like:
                      bankbankincountry(royal, ireland)
Here, royal refers to a different real-world entity, namely the Royal Bank of
Scotland. Hence, it is important to uniquely identify the terms in accordance
with the contextual information provided by the entire triple. In this paper, we
aim at aligning such polysemous terms from open IE systems to instances from
a closed IE system, while focusing on NELL and DBpedia in particular.
    The remainder of the paper is organized as follows, in Section 2 we introduce
the information extraction projects relevant to our work. We present our baseline
algorithm for finding the best matching candidates for a term in Section 3 and in
Section 4 introduce a gold standard for evaluating its performance. In Section 5
we report performance results of the proposed approach. In Section 6 we discuss
related work on information extraction and entity linking. Finally, we conclude
the paper in Section 7.


2     Information Extraction Projects: A Brief Overview

The Never Ending Language Learning [5] (Nell) project’s objective is the
creation and maintenance of a large-scale machine learning system that continu-
ously learns and extracts structured information from unstructured web pages.
Its extraction algorithms operate on a large corpus of more than 500 million
web pages1 and not solely on the set of Wikipedia articles. The NELL sys-
tem was bootstrapped with a small set of classes and relations and, for each
of those, 10-15 positive and negative instances. The guiding principle of NELL
is to build several semi-supervised machine learning [6] components that ac-
cumulate instances of the classes and relations, re-train the machine learning
algorithms with these instances as training data, and re-apply the algorithms to
1
    http://lemurproject.org/clueweb09/
                                    Linking Information Extraction Projects       3

extract novel instances. This process is repeated indefinitely with each re-training
and extraction phase called an iteration. Since numerous extraction components
work in parallel and extract facts with different degrees of confidence in their
correctness, one of the most important aspects of Nell is its ability to combine
these different extraction algorithms into one coherent model. This is also ac-
complished with relatively simple linear machine learning algorithms that weigh
the different components based on their past accuracy.
    Nell has been running since 2010, initially fully automated and without
any human supervision. Since it has experienced concepts drift for some of its
relations and classes, that is, an increasingly worse extraction performance over
time, Nell now is given some corrections by humans to avoid this long-term
behavior. Nell does not adhere to any of the semantic web standards such as
RDF or description logic.

DBpedia [1, 17] is a project that aims at automatically acquiring large amounts
of structured information from Wikipedia. It extracts information from infobox
templates, categories, geo-coordinates, etc.. However, it does not learn relations
from the Wikipedia categories. This template information is mapped to an on-
tology. In addition, it has a fixed set of classes and relations. Moreover, the
ontology is with more than 1000 different relations much broader than other
existing ontologies like YAGO [25] or semantic lexicons like BabelNet [19].
    DBpedia represents its data in accordance with the best-practices of pub-
lishing linked open data. The term linked data describes an assortment of best
practices for publishing, sharing, and connecting structured data and knowledge
over the web [2]. DBpedia’s relations are modeled using the resource descrip-
tion framework (RDF), a generic graph-based data model for describing objects
and their relationships. The entities in DBpedia have unique URIs. This makes
it appropriate as our reference knowledge base to which we can link the terms
from Nell. In the case of the examples from Section 1, by linking the terms
appropriately to DBpedia, we are able to attach an unambiguous identifier to
them which was initially missing.
       royals ⇒ http://dbpedia.org/resource/Kansas City Royals
   royal ⇒ http://dbpedia.org/resource/The Royal Bank of Scotland


3   Methodology

Wikipedia is an exhaustive source of unstructured data which has been exten-
sively used to enrich machines with knowledge [15]. In this work we use Wikipedia
as an entity-tagged corpus [4] in order to bridge knowledge encoded in Nell with
DBpedia. Since there is a corresponding DBpedia entity for each Wikipedia
article [2], we can in fact formulate our disambiguation problem as that of link-
ing entities mentioned within Nell triples to their respective Wikipedia articles.
Our problem is that, due to polysemy, often a term from Nell can refer to sev-
eral different articles in Wikipedia or, analogously, instances in DBpedia. For
4       A. Dutta et al.

instance, the term jaguar can refer to several articles such as the car, the animal
and so on.
    In this work we accordingly ex- anchor               Article        Link count
plore the idea of using Wikipedia to jaguar          Jaguar Cars           1842
find out the most probable article for jaguar Jaguar Racing                 440
a given term. Wikipedia provides reg- jaguar             Jaguar             414
ular data dumps and there are off-
                                           ...             ...              ...
the-shelf preprocessing tools to parse
those dumps. We used WikiPrep [11, lincoln Lincoln, England                1844
10] for our purpose. WikiPrep re-        lincoln  Lincoln,    Nebraska      920
moves redundant information from         lincoln Lincoln   (2012  film)     496
the original dumps and creates more        . . .           . . .            ...
relevant XML dumps with additional
                                                 Table 1. Snippet of those articles linked to
information like the number of pages using the anchor jaguar and lincoln.
in each category, incoming links to
each Wikipedia article and their anchor text, and a lot more2 . In our work, we
are primarily interested in the link counts, namely the frequency of anchor text
labels pointing to the same Wikipedia page. Table 1 shows some of the articles
the anchors jaguar or lincoln are referring to. Intuitively, out of all the outgoing
links from the anchor term jaguar, 1842 links pointed to the article Jaguar Cars
and so on. Essentially, these anchors are analogous to the NELL terms. Based
on these counts, we create a ranked list of articles for a given anchor3 .
    As seen in Table 1, the output from WikiPrep can often be a long list of
anchor-article pairs and some of them having as low as just one link count.
Accordingly, we adopt a probabilistic approach in selecting the best possible
DBpedia instance. For any given anchor in Wikipedia, the fraction of articles
the links points to is proportional to the probability that the anchor term refers
to the particular article [23]. More formally, suppose some anchor e refers to N
articles A1 , . . . , AN with n1 , . . . , nN respective links counts,Pthen the conditional
                                                                        N
probability P of e referring to Aj is given by, P (Aj |e) = nj / i=1 ni . We compute
the probabilities for each terms we are interested in and from the ranked list of
descending P (Aj |e), top-k candidates are selected. The choice of k is described
in Section 4. We apply this idea on the Nell data set. For each Nell triple, we
take the terms occurring as subject and object, and apply the procedure above.


4    Creating a Gold Standard
Nell provides regular data dumps4 consisting of facts learned from the Web.
Based on this data we create a frequency distribution over the predicates. To this
end, we first clean up the data from the dumps (since these contain additional
2
  http://www.cs.technion.ac.il/~gabr/resources/code/wikiprep/
3
  Note that, while there are alternative data sets such as the Crosswiki data [23], in
  this work we opted instead for exploiting only Wikipedia internal-link anchors since
  we expect them to provide a cleaner source of data.
4
  http://rtw.ml.cmu.edu/rtw/resources
                                        Linking Information Extraction Projects             5

        Top predicates           Instances          Random predicates               Instances
         generalizations          1297709        personleads-organization              716
             proxyfor               5540     countrylocatedingeopoliticallocation      632
          agentcreated              4354            actorstarredinmovie                537
            subpartof               3262           athleteledsportsteam                294
           atlocation               2877               personbornincity                285
         mutualproxyfor             2803            bankbankincountry                  246
 locationlocatedwithinlocation      2159          weaponmadeincountry                  188
        athleteplayssport           2076              athletebeatathlete               148
      citylocatedinstate            2010           companyalsoknownas                  107
  professionistypeofprofession      1936                 lakeinstate                   105
     subpartoforganization          1874
          bookwriter                1809
     furniturefoundinroom           1674
agentcollaborateswithagent          1541
    animalistypeofanimal            1540
      agentactsinlocation           1490
   teamplaysagainstteam             1448
      athleteplaysinleague          1390
             worksfor               1303
   chemicalistypeofchemical         1303

Table 2. The 30 most frequent predicates found in Nell. The set of predicates we
randomly sampled for the gold standard are in bold.




information, such as, for instance, iteration of promotion, best literal strings, and
so on5 , which are irrelevant to our task). In Table 2, we list the 30 most frequent
predicates. Since the gold standard should not be biased towards predicates with
many assertions we randomly sampled 12 predicates from the set of predicates
with at least 100 assertions (highlighted in bold in the table). In this paper, we
focus on this smaller set of predicates due to the time consuming nature of the
manual annotations we needed to perform. However, we plan to continuously
extend the gold standard with additional predicates in the future.
    For each Nell predicate we randomly sampled 100 triples. We assigned each
predicate and the corresponding list of triples to an annotator. Since we wanted
to annotate a large number of triples within an acceptable time frame, we first
applied the method described in Section 3 to generate possible mapping candi-
dates for the Nell subject and object of each triple. In particular, we generated
the top-3 mappings, thereby avoiding generation of too many possible candi-
dates, and presented those candidates to the annotator. Note that in some cases
(see Table 3), our method could not determine a possible mapping candidate
for a Nell instance. In this case, the triple had to be annotated without pre-
senting a matching candidate for subject or object or both. In our setting, each
annotation instance falls under one of the following three cases:

 (i) One of the mapping candidates is chosen as the correct mapping, i.e., the
     simplest case.
(ii) The correct mapping is not among the presented candidates (or no candi-
     dates have been generated). However, the annotator can find the correct
5
    http://rtw.ml.cmu.edu/rtw/faq
6         A. Dutta et al.

 Nell-Subject       Nell-Object              DBP-Subject                  DBP-Object
   stranger         albert-camus          The Stranger (novel)            Albert Camus
                      1st cand.             Stranger (comics)             Albert Camus
                      2nd cand.        Stranger (Hilary Duff song)               -
                      3rd cand.            Characters of Myst                    -
      gospel        henry james                     ?                      Henry James
                      1st cand.               Gospel music                 Henry James
                      2nd cand.                  Gospel              Henry James (basketball)
                      3rd cand.        Urban contemporary gospel     Henry James, 1st Baron...
 riddle master   patricia a mckillip    The Riddle-Master of Hed       Patricia A. McKillip
                      1st cand.                     -                  Patricia A. McKillip
                      2nd cand.                     -                            -
                      3rd cand.                     -                            -
    king john       shakespeare             King John (play)           William Shakespeare
                      1st cand.          John, King of England         William Shakespeare
                      2nd cand.             King John (play)          Shakespeare quadrangle
                      3rd cand.           King John (1899 film)        Shakespeare, Ontario

Table 3. Four annotation examples of the bookwriter predicate (we have removed the
URI prefix http://dbpedia.org/resource/ for better readability).



      mapping after a combined search in DBpedia, Wikipedia or other resources
      available on the Web.
(iii) The annotator cannot determine a DBpedia entity to which the given Nell
      instance should be mapped. This was the case when the term was too am-
      biguous, underspecified, or not represented in DBpedia. In this case the
      annotator marked the instance as unmatchable (‘?’).
   Table 3 shows four possible annotation outcomes for the bookwriter predi-
cate. The first example illustrates case (i) and (ii). The second example illustrates
case (i) and (iii). With respect to this example the annotator could not determine
the reference for the Nell term gospel. The third example illustrates a special
case of (ii) where no mapping candidate has been generated for the Nell term
patricia a mckillip. The fourth example shows that the top match generated
by our algorithm is not always the correct mapping, but might also be among
the other alternatives that have been generated.


5      Experiments
5.1      Evaluation Measures
In the following, we briefly re-visit the definitions of precision and recall and
explain their application in our evaluation scenario. Let A refer to the mappings
generated by our algorithm, and G refer to mappings in the gold standard.
Precision is defined as prec(A, G) = |A ∩ G|/|A| and recall as rec(A, G) =
|A ∩ G|/|G|. The F1 measure is the equally weighted harmonic mean of both
values, i.e., F1 (A, G) = 2 ∗ prec(A, G) ∗ rec(A, G)/(prec(A, G) + rec(A, G)).
   If an annotator assigned a question mark, then the corresponding Nell term
could not be mapped and it does not appear in the gold standard G. This
can again be seen in Table 3, where we present the mappings generated by
                                    Linking Information Extraction Projects                7

         Predicate           Unmatched        Reason/Observation
agentcollaborateswithagent      15%           only first names or surnames are given
         lakeinstate            14%           non-existent entities for the lakes in context
 personleadsorganization       9.5%           non-existent entities or too ambiguous
         bookwriter            8.5%           obscure books, writer ambiguous
   animalistypeofanimal          8%           uncommon description of animal types
  teamplaysagainstteam         6.5%           ambiguity between college and college team
   companyasloknownas          5.5%           ambiguous names
  weaponmadeincountry          3.5%           non-existent entities in DBpedia
    actorstarredinmovie          3%           ambiguity between film, acts, or play
    bankbankincountry            2%           ambiguous and non-existent entities
     citylocatedinstate        0.5%           too general entity
   athleteledsportsteam          0%           well defined names of persons

Table 4. The percentage of entities per predicate that could not be matched by a
human annotator.




our algorithm for four triples, as well as the corresponding gold-standard an-
notations. If the mapping A consists of top-k possible candidates, computing
precision and recall on the examples, we have the precision value for k = 1 as
prec@1 = 4/7 ≈ 57% and rec@1 = 4/7 ≈ 57%. Note that precision and recall are
not the same in general, because |A| =6 |G| in most cases. More generally, we are
interested in prec@k, the fraction of top-k candidates that are correctly mapped
and rec@k, the fraction of correct mappings that are in the top-k candidates.
For k = 3, we have prec@3 = 5/17 ≈ 29% and rec@3 = 5/7 ≈ 71%.
    It can be expected that prec@1 will have the highest score and rec@1 will
have the lowest score. When we analyze A with k > 1, we focus mainly on the
increase in recall. Here we are in particular interested in the value of k for which
the number of additionally generated correct mappings in A is negligibly small
compared to the mappings generated in A for k + 1.
    When generating the gold standard, we realized that finding the correct map-
pings is often a hard task and sometimes even difficult for a human annotator.
We had also observed that the problem of determining the gold standard varies
strongly across the properties we analyzed. For some of the properties we could
match all (or nearly all) subjects and objects in the chosen triples, while for
other properties up to 15% of the instances could not be matched.
   Table 4 presents the percentage of entities that could not be matched by
the annotators, together with the main reason the annotators provided when
they could not find a corresponding entity in DBpedia. A typical example of a
problematic triple from the agentcollaborateswithagent property is
                   agentcollaborateswithagent(world, greg)

In this case, the mapping for subject and object was annotated with a ques-
tion mark. We also observed cases in which an uncommon description was cho-
sen that had no counterpart in DBpedia. Some examples from the predicate
animalistypeofanimal are the labels furbearers or small mammals.
8                                  A. Dutta et al.


                                                                            precision@1                                         recall@1


                   1.0
                                                                                               0.890.86                                                                                                               0.880.88        0.9 0.92
                                                                 0.86        0.860.88
                                       0.810.82           0.82                                                    0.82                0.8 0.81     0.820.83                                          0.810.81
                   0.8                                                                                                   0.78                                                       0.79
                                                                                                                                                                   0.71                    0.75
Precision/Recall




                                                                                                                                                                          0.58
                   0.6
                   0.4
                   0.2
                   0.0




                                                                                   l                                                                    r                                                                                   te
                                       mov
                                            ie
                                                          agen
                                                               t
                                                                              nima             stea
                                                                                                    m            untry           dins
                                                                                                                                      tate        write             nas           izatio
                                                                                                                                                                                         n            eam         coun
                                                                                                                                                                                                                       try          insta
                                 edin               swith               peofa             sport             inco            cate             book              know          rgan                instt
                                                                                                                                                                                                            adein            lake
                          rstarr              orate              alisty             teled              bank          citylo                               aslo          adso                saga
                   acto                 ollab              anim               athle               bank                                              pany           onle                 play           ponm
                                 e n tc                                                                                                        co m           pers                team             wea
                              ag


                                                                                                                                      NELL Properties



                                                            Fig. 1. prec@1 and rec@1 of our proposed method.


5.2                           Results and Discussion
We run our algorithm against the gold standard6 , and report the precision and
recall values. In Figure 1, we show the precision and recall values obtained on
the set of Nell predicates. These values are for top-1 matches. Precision and
recall vary across the predicates with lakeinstate having the highest precision.
Using micro-average method, for the top-1 matches we achieved a precision of
82.78% and an average recall of 81.31% across all the predicates. In the case of
macro-averaging, instead, we achieved precision of 82.61% and recall of 81.42%.
   In Figure 2, we show the values for rec@2, rec@5 and rec@10 compared
to rec@1, the recall values reported in Figure 1. By considering more possible
candidates with increasing k, every term gets a better chance of being matched
correctly, thus explaining the increases in rec@k with k. However, it must be
noted, that for most of the predicates the values tend to saturate after rec@5.
This reflects that after a certain k any further increase in k does not alter the
correct mappings, since our algorithm already provided a match within top-1 or
top-2 candidates. Still, for some we observe an increase even at rec@10 because
there can be still a possibility of one correct matching candidate lying at a much
lower rank in the top-k list of candidates.
   In Figure 3, we plot the micro-average values of the precision, recall and F1
scores over varying k. We attain the best F1 score of 0.82 for k = 1 and the
recall values tend to saturate after k = 5.
   This raises an important question regarding the upper bound of recall of our
algorithm. In practice, we cannot achieve a recall of 1.0 because we are limited
from factors like:
             – the matching candidate being never referred to by the terms. For example,
               gs refers to the company Goldmann Sachs, but it never appeared even in all
               the possible candidates, since Goldmann Sachs is never referred to with gs
               in Wikipedia.
    6
                   The data are freely available at http://web.informatik.uni-mannheim.de/data/
                   nell-dbpedia/NellGoldStandard.tar.
                                                                                                                                                                  Linking Information Extraction Projects                                                                                                                                                                         9


                                                                                                                                       rec@1                                                              rec@2                                           rec@5                                                        rec@10



         1.0
         0.8
         0.6
Recall

         0.4
         0.2
         0.0




                                                                                                  al                                                                                                 te                      r                                                                                                                                               te
                                    ov
                                      ie                           nt                                                              m                              try                          sta                     wr
                                                                                                                                                                                                                         ite                       as                                  n                           m                               try                 sta
                                                              ge                             nim                             tea                             un                                                   ok                          wn                             iza
                                                                                                                                                                                                                                                                                   tio
                                                                                                                                                                                                                                                                                                           stt
                                                                                                                                                                                                                                                                                                              ea                              un                 ein
                              inm                        ha                            ofa                             rts                              co                               din                 bo                          no                             an                                                               co                lak
                         ed                           wit                          e                            sp
                                                                                                                   o                              kin                              ate                                                lok                                                            ain                           ein
                      arr                           es                          yp                          led                            ba
                                                                                                                                              n                                c                                                   as                            o rg                           ag                             d
                   rst                           rat                    a   list                       te                             n   k                             cit
                                                                                                                                                                           ylo                                                   ny                        a   ds                          ys                           n   ma
           ac
              to                          ab
                                             o                      im                            thle                             ba                                                                                    mp
                                                                                                                                                                                                                            a
                                                                                                                                                                                                                                                       nle                           pla                             po
                                 tco
                                     ll                        an                             a                                                                                                                    co                            rso                          te   am                       we
                                                                                                                                                                                                                                                                                                                 a
                            en                                                                                                                                                                                                              pe
                       ag
                                                                                                                                                                                               NELL Predicates



                                                                                Fig. 2. Comparison of rec@1 against rec@k.


     – persons being often referred to by the combination of their middle and last
       name. For e.g. hussein obama. It is actually talking about President Barack
       Obama, but with our approach we cannot find a good match.
     – misspelled words. We have entities like missle instead of missile.

However, there are ways to further improve the recall of our method like, for in-
stance, by means of string similarity techniques – e.g., Levenshtein edit distance.
A similarity threshold (say, as high as 95%) could then be tuned to consider en-
tities which only partially match a given term. Another alternative would be
to look for sub-string matches for the terms with middle and last names of
persons. For instance, hussein obama can have a possible match if terms like
barrack hussein obama has a candidate match. In addition, a similarity thresh-
old can be introduced in order to avoid matching by arbitrary longer terms.
    In general, thanks to the an-
notation task and our experi-
ments we were able to acquire
                                                                                                                                                                                     1.0




some useful insights about the
data set and the proposed task.                 0.83                       0.89
                                             0.82        0.86    0.89
                                                                                                                                                                                     0.8




                                                                                                                                                                                                                        0.81

     – Predicates with polysemous                                                                                                                                                                                                                      0.63
                                                                                                                                                                                     0.6




       entities, like companyalso
                                                                                                                                                                          Scores




                                                                                                                                                                                                                                                                    0.49
       knownas, usually have lower
                                                                                                                                                                                     0.4




       precision. The triples for this                                                                                                                                                                                                                                                           0.38

       predicate had a wide usage of                                                                                                                                                                                                                                                                         0.24
                                                                                                                                                                                                                                                                                                                                               0.24
                                                                                                                                                                                     0.2




       abbreviated terms (the stock                                                                                                                                                                                                                                                                                                               0.14
       exchange codes for the com-                                                                                                                                                                                precision@k
                                                                                                                                                                                                                  recall@k
       panies) and that accounts for
                                                                                                                                                                                     0.0




                                                                                                                                                                                                                  F1
       a lower precision value.                                                                                                                                                                                         k=1                                     k=2                                         k=5                                          k=10                         k=1

     – The Nell data is skewed to-                                                                                                                                                                                                                                           top-k values

       wards a particular region or
       type. The triples involving                                                                                                                                          Fig. 3. Micro-average prec@k, rec@k and F1 .
10      A. Dutta et al.

     persons and sports primarily refer to basketball or baseball. Similarly, for
     lakeinstate, nearly all the triples refer to lakes in United States.


6    Related Work

Key contributions in information extraction have concentrated on minimizing
the amount of human supervision required in the knowledge harvesting process.
To this end, much work has explored unsupervised bootstrapping for a variety
of tasks, including the acquisition of binary relations [3], facts [8], semantic
class attributes and instances [20]. Open Information Extraction further focused
on approaches that do not need any manually-labeled data [9], however, the
output of these systems still needs to be disambiguated by linking it to entities
and relations from a knowledge base. Recent work has extensively explored the
usage of distant supervision for IE, namely by harvesting sentences containing
concepts whose relation is known and leveraging these sentences as training data
for supervised extractors [27, 14]. Talking of integration of open and closed IE
projects, it is worthwhile to mention the work of [21] where matrix factorization
technique was employed for extracting relations across different domains. They
proposed an universal schema which supports cross domain integration.
    There has been some work on instance matching in the recent past. Re-
searchers have transformed the task into a binary classification problem and
solved it with machine learning techniques [22]. Some have tried to enrich un-
structured data in form of text with Wikipedia entities [18]. However, in our
approach we consider the context of the entities while creating the gold stan-
dard which makes it bit different from these above mentioned entity linking
approaches. Also, there are tools like Tı̀palo [12] for automatic typing of DB-
pedia entities. They use language definitions from Wikipedia abstracts and use
WordNet in the background for disambiguation. PARIS [24] takes a probabilis-
tic approach to align ontologies utilizes the interdependence of instances and
schema to compute probabilities for the instance matches. Lin et al. [16] pro-
vide a novel approach to link entities across million documents. They take web
extracted facts and link the entities to Wikipedia by means of information from
Wikipedia itself, as well as additional features like string similarity, and most
importantly context information of the extracted facts. The Silk framework [26]
discovers missing links between entities across linked data sources by employing
similarity metrics between pairs of instances.


7    Conclusions

In this paper, we introduced a most-frequent-entity baseline algorithm in order
to link entities from an open domain system to a closed one. We introduced a
gold standard for this task and compared our baseline against it. In the near
future, we plan to extend this work with more complex and robust methods, as
well as extending our methodology to cover other open IE projects like ReVerb.
                                     Linking Information Extraction Projects        11

References
 1. Sören Auer, Christian Bizer, Georgi Kobilarov, Jens Lehmann, Richard Cyganiak,
    and Zachary Ive. Dbpedia: A nucleus for a web of open data. In Proceedings of
    6th International Semantic Web Conference joint with 2nd Asian Semantic Web
    Conference (ISWC+ASWC 2007), pages 722–735, 2007.
 2. Christian Bizer, Jens Lehmann, Georgi Kobilarov, Sören Auer, Christian Becker,
    Richard Cyganiak, and Sebastian Hellmann. DBpedia – A crystallization point for
    the web of data. Journal of Web Semantics, 7(3):154–165, 2009.
 3. Sergey Brin. Extracting patterns and relations from the world wide web. In Selected
    papers from the International Workshop on The World Wide Web and Databases,
    WebDB ’98, pages 172–183. Springer-Verlag, 1999.
 4. Razvan Bunescu and Marius Paşca. Using encyclopedic knowledge for named entity
    disambiguation. In Proc. of EACL-06, pages 9–16, 2006.
 5. Andrew Carlson, Justin Betteridge, Bryan Kisiel, Burr Settles, Estevam R. Hr-
    uschka, and Tom M. Mitchell. Toward an architecture for never-ending language
    learning. In Proc. of AAAI-10, pages 1306–1313, 2010.
 6. Olivier Chapelle, Bernhard Schlkopf, and Alexander Zien. Semi-Supervised Learn-
    ing. MIT Press, 1st edition, 2010.
 7. Oren Etzioni. Search needs a shake-up. Nature, 476(7358):25–26, 2011.
 8. Oren Etzioni, Michael Cafarella, Doug Downey, Stanley Kok, Ana-Maria Popescu,
    Tal Shaked, Stephen Soderland, Daniel S. Weld, and Alexander Yates. Web-scale
    information extraction in KnowItAll: (preliminary results). In Proc. of WWW ’04,
    pages 100–110, 2004.
 9. Anthony Fader, Stephen Soderland, and Oren Etzioni. Identifying relations for
    open information extraction. In Proc. of EMNLP-11, pages 1535–1545, 2011.
10. Evgeniy Gabrilovich and Shaul Markovitch. Overcoming the brittleness bottleneck
    using wikipedia: Enhancing text categorization with encyclopedic knowledge. In
    Proc. of AAAI-06, pages 1301–1306, 2006.
11. Evgeniy Gabrilovich and Shaul Markovitch. Computing semantic relatedness using
    wikipedia-based explicit semantic analysis. In Proc. of IJCAI-07, pages 1606–1611,
    2007.
12. Aldo Gangemi, AndreaGiovanni Nuzzolese, Valentina Presutti, Francesco Draic-
    chio, Alberto Musetti, and Paolo Ciancarini. Automatic typing of DBpedia entities.
    In The Semantic Web – ISWC 2012, volume 7649 of Lecture Notes in Computer
    Science, pages 65–81. Springer, Berlin and Heidelberg, 2012.
13. Johannes Hoffart, Fabian M. Suchanek, Klaus Berberich, and Gerhard Weikum.
    YAGO2: A spatially and temporally enhanced knowledge base from Wikipedia.
    Artificial Intelligence, 194:28–61, 2012.
14. Raphael Hoffmann, Congle Zhang, and Daniel S. Weld. Learning 5000 relational
    extractors. In Proc. of ACL-10, pages 286–295, 2010.
15. Eduard Hovy, Roberto Navigli, and Simone Paolo Ponzetto. Collaboratively built
    semi-structured content and Artificial Intelligence: The story so far. Artificial
    Intelligence, 194:2–27, 2013.
16. Thomas Lin, Mausam, and Oren Etzioni. Entity linking at web scale. In Proc. of
    AKBC-WEKEX ’12, pages 84–88, 2012.
17. Pablo Mendes, Max Jakob, and Christian Bizer. Dbpedia: A multilingual cross-
    domain knowledge base. In Proc. of LREC-12, 2012.
18. David Milne and Ian H. Witten. Learning to link with wikipedia. In Proc. of
    CIKM ’08, pages 509–518, 2008.
12      A. Dutta et al.

19. Roberto Navigli and Simone Paolo Ponzetto. BabelNet: The automatic construc-
    tion, evaluation and application of a wide-coverage multilingual semantic network.
    Artificial Intelligence, 193:217–250, 2012.
20. Marius Paşca. Organizing and searching the world wide web of facts – step two:
    harnessing the wisdom of the crowds. In Proc. of WWW ’07, pages 101–110, 2007.
21. Sebastian Riedel, Limin Yao, Benjamin M. Marlin, and Andrew McCallum. Rela-
    tion extraction with matrix factorization and universal schemas. In Joint Human
    Language Technology Conference/Annual Meeting of the North American Chapter
    of the Association for Computational Linguistics (HLT-NAACL ’13), 2013.
22. Shu Rong, Xing Niu, EvanWei Xiang, Haofen Wang, Qiang Yang, and Yong Yu.
    A machine learning approach for instance matching based on similarity metrics.
    In The Semantic Web – ISWC 2012, volume 7649 of Lecture Notes in Computer
    Science, pages 460–475. Springer, Berlin and Heidelberg, 2012.
23. Valentin I. Spitkovsky and Angel X. Chang. A cross-lingual dictionary for english
    wikipedia concepts. In Proc of LREC-12, pages 3168–3175, 2012.
24. Fabian M. Suchanek, Serge Abiteboul, and Pierre Senellart. Paris: probabilistic
    alignment of relations, instances, and schema. Proc. VLDB Endow., 5(3):157–168,
    November 2011.
25. Fabian M. Suchanek, Gjergji Kasneci, and Gerhard Weikum. Yago: A Core of
    Semantic Knowledge. In 16th international World Wide Web conference (WWW
    2007), New York, NY, USA, 2007. ACM Press.
26. Julius Volz, Christian Bizer, Martin Gaedke, and Georgi Kobilarov. Silk - A Link
    Discovery Framework for the Web of Data. In Proc. of LDOW ’09, 2009.
27. Fei Wu and Daniel S. Weld. Open information extraction using Wikipedia. In
    Proc. of ACL-10, pages 118–127, 2010.