<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>Speci cally we used Wikipedia</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>End-to-End Learning for Answering Structured Queries Directly over Text</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Paul Groth</string-name>
          <email>p.groth@springer.com</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Antony Scerri</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ron Daniel</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Bradley P. Allen</string-name>
          <email>b.alleng@elsevier.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Elsevier Labs</institution>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>University of Amsterdam</institution>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2018</year>
      </pub-date>
      <volume>201</volume>
      <fpage>8</fpage>
      <lpage>08</lpage>
      <abstract>
        <p>Structured queries expressed in languages (such as SQL, SPARQL, or XQuery) o er a convenient and explicit way for users to express their information needs for a number of tasks. In this work, we present an approach to answer these directly over text data without storing results in a database. We speci cally look at the case of knowledge bases where queries are over entities and the relations between them. Our approach combines distributed query answering (e.g. Triple Pattern Fragments) with models built for extractive question answering. Importantly, by applying distributed querying answering we are able to simplify the model learning problem. We train models for a large portion (572) of the relations within Wikidata and achieve an average 0.70 F1 measure across all models. We describe both a method to construct the necessary training data for this task from knowledge graphs as well as a prototype implementation.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>Database query languages (e.g. SQL, SPARQL, XQuery) o er a convenient and
explicit way for users to express their information needs for a number of tasks
including populating a dataframe for statistical analysis, selecting data for display
on a website, de ning an aggregation of two datasets, or generating reports.</p>
      <p>
        However, much of the information that a user might wish to access using a
structured query may not be available in a database and instead available only
in an unstructured form (e.g. text documents). To overcome this gap, the area
of information extraction (IE) speci cally investigates the creation of structured
data from unstructured content [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ]. Typically, IE systems are organized as
pipelines taking in documents and generating various forms of structured data
from it. This includes the extraction of relations, the recognition of entities, and
even the complete construction of databases. The goal then of IE is not to answer
queries directly but rst to generate a database that queries can be subsequently
executed over.
      </p>
      <p>
        In the mid-2000s, with the rise of large scale web text, the notion of
combining information extraction techniques with relational database management
systems emerged [
        <xref ref-type="bibr" rid="ref10 ref4">4, 10</xref>
        ] resulting in what are termed text databases. Systems
like Deep Dive [
        <xref ref-type="bibr" rid="ref22">22</xref>
        ] InstaRead [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ], or Indrex [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ], use database optimizations
within tasks such as query planning to help decide when to perform extractions.
While, in some cases, extraction of data can be performed at runtime, data is
still extracted to an intermediate database before the query is answered. Thus,
all these approaches still require the existence of a structured database to answer
the query.
      </p>
      <p>
        In this paper, we present an approach that eliminates the need to have an
intermediate database in order to answer structured database queries
over text. This is essentially the same as treating the text itself as the store of
structured data. Using text as the database has a number of potential bene ts,
including being able to run structured queries over new text without the need for
a-priori extraction; removing the need to maintain two stores for the same
information (i.e. a database and a search index); eliminating synchronization issues;
and reducing the need for up-front schema modeling. [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] provides additional
rationale for not pre-indexing \raw data", although they focus on structured data
in the form of CSV les.
      </p>
      <p>
        Our approach builds upon three foundations: 1. the existence of large scale
publicly available knowledge bases (Wikidata) derived from text data (Wikipedia);
2. recent advances in end-to-end learning for extractive question answering (e.g.
[
        <xref ref-type="bibr" rid="ref21">21</xref>
        ]); 3. the availability of layered query processing engines designed for
distributed data (e.g. SPARQL query processes that work over Triple Pattern
Fragment [
        <xref ref-type="bibr" rid="ref25">25</xref>
        ] servers).
      </p>
      <p>A high-level summary of our approach is as follows. We use a
publiclyavailable knowledge base to construct a parallel corpus consisting of tuples each
which is made up of a structured slot lling query, the expected answer drawn
from the knowledge base, and a corresponding text document in which we know
the answer is contained. Using this corpus, we train neural models that learn to
answer the given structured query given a text document. This is done on a per
relation basis. These models are trained end-to-end with no speci c tuning for
each query. These models are integrated into a system that answers queries
expressed in a graph query language directly over text with no relational or graph
database intermediary.</p>
      <p>The contributions of this paper are:
{ an approach, including training data generation, for the task of answering
structured queries over text;
{ models that can answer slot lling queries for over 570 relations with no
relation or type speci c tuning. These models obtain on average a 0.70 F1
measure for query answering.
{ a prototype system that answers structured queries using triple pattern
fragments over a large corpus of text (Wikipedia).</p>
      <p>The rest of this paper is organized as follows. We begin with an overview of
the approach. This is followed by a description of the training data. Subsequently,
we describe the model training and discuss the experimental results. After which,
we present our prototype system. We end the paper with a discussion of related
and future work.</p>
    </sec>
    <sec id="sec-2">
      <title>Overview</title>
      <p>SPARQL TPF Client</p>
      <sec id="sec-2-1">
        <title>TPF Q(FuaecraydEen)gine</title>
        <p>QA Executor
Answer Ranking</p>
        <p>Model Runner
Model Query Builder
&lt;s, p, o&gt; to string
Candidate</p>
        <p>Selection
Mo1del …</p>
      </sec>
      <sec id="sec-2-2">
        <title>MoNdel</title>
        <p>Query Service</p>
        <p>Text
Search Engine
Document
Col ection
Col</p>
        <p>
          Our overall approach consists of several components as illustrated in Figure
1. First, structured queries as expressed by SPARQL [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ] are executed using a
Triple Pattern Client (SPARQL TPF Client). Such a client breaks down a more
complex SPARQL query into a series of triple patterns that are then issued to a
service. Triple patterns are queries of the form subject, predicate, object, where
each portion can be bound to an identi er (i.e. URI) or a variable.3 Within the
service, the execution engine (QA Executor) rst lexicalizes the given identi ers
into strings using the Model Query Builder component. For example, this
component would translate an identi er like https://www.wikidata.org/wiki/Q727
into the string form \Amsterdam". These queries are then issued to a candidate
selection component. This component queries a standard text search engine to
nd potential documents that could contain the answer to the speci ed query.
        </p>
        <p>The candidate documents along with the lexicalized queries are provided
to a model runner which issues these to models trained speci cally to bind
the variable that is missing. That is given a query of the form &lt; s; p; ?o &gt;
where s and o are bound and o is the variable, there would be speci c models
trained to extract ?o from the provided document. For example, given the query
(:Amsterdam :capital of ?o) we would have models that know how to answer
queries of where the type of the subject is City and the property is capital of.
Likewise, there would be models of that are able to answer queries of the form
&lt;?s; p; o &gt; and so on. Each model is then asked to generate a binding of the
variable. Note that the bindings generated by the models are strings. The results
of each model are then ranked (Answer Ranking). Using a cut-o , the results
are then translated back into identi er space and returned to the client.</p>
        <sec id="sec-2-2-1">
          <title>3 Objects can also be bound to a literal.</title>
          <p>A key insight of our approach is that by breaking down complex queries
into triple patterns we can simplify the queries that need to be answered by the
learned models.</p>
          <p>
            Our approach relies on the construction of models that are able to extract
potential candidate answers from text. Following from [
            <xref ref-type="bibr" rid="ref5">5</xref>
            ] and [
            <xref ref-type="bibr" rid="ref13">13</xref>
            ], we cast the
problem in terms of a question answering task, where the input is a question (e.g.
entity type + relation) and a document and the output is answer span within
the document that binds the output. To learn these sorts of models we construct
training data from knowledge graphs that have a corresponding representation
in text. In the next section, we go into detail about the construction of the
necessary training data.
3
          </p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Training Data Construction</title>
      <p>
        Our training data is based on the combination of Wikidata and Wikipedia.
Wikidata is a publicly accessible and maintained knowledge base of encyclopedic
information [
        <xref ref-type="bibr" rid="ref27">27</xref>
        ]. It is a graph structured knowledge base (i.e. a knowledge graph)
describing entities and the relations between them. Every entity has a globally
unique identi er. Entities may also have attributes which have speci c datatypes.
Entities have may have more than one type. Relations between entities may hold
between di ering entity types.
      </p>
      <p>Wikidata has a number of properties that make it useful for building a corpus
to learn how to answer structured queries over text. First, and perhaps most
importantly, entities have a parallel description with Wikipedia. By our count,
Wikidata references 7.7 million articles in the English language Wikipedia. Thus,
we have body of text which will also most likely contain answers that we retrieve
from Wikidata. Second, every entity and relation in Wikidata has a human
readable label in multiple languages. This enables us to build a direct connection
between the database and text. Third, Wikidata is large enough to provide for
adequate training data in order to build models. Finally, Wikidata provides
access to their data in a number of ways including as a SPARQL endpoint, a
triple patterns fragment endpoint and as a bulk RDF download. While we use
Wikidata, we believe that our approach can be extended to any knowledge graph
that has textual sources.</p>
      <p>Using this input data we generate datasets of the form: [query; answer;
text in which the query is answered]. As previously mentioned,
complex queries can be expressed as a series of graph patterns. Thus, the queries
we consider are graph patterns in which two of the variables are bound (e.g.
:New England Patriots :play ?x). We term these slot lling queries as the
aim is to bind one slot in the relation (i.e. the subject or the object). While we do
not test graph patterns where the predicate is the variable, the same approach
is also applicable. In some sense, one can think of this as generating data that
can be used to build models that act as substitute indexes of a database.</p>
      <p>Our construction method loops through all of the predicates (i.e. relations) in
the dataset. It determines the frequency with which a predicate connects di erent
types of entities. This is essential as large knowledge graphs can connect many
di erent types using the same predicate. Thus, examples from di erent types
of subjects and objects are needed to capture the semantics of that predicate.
Using the most frequently occurring pairs of entity types for a predicate, the
algorithm then retrieves as many example triples as possible where the subject
and object of the triple are instances of the connected types - up to a given
maximum threshold. Thresholding is used to help control the size of the training
data.</p>
      <p>Each triple is then used to generate a row of training data for learning how
to answer graph pattern queries that contain the given predicate. To connect
the graph pattern queries, which are expressed using entity IRIs to the plain
text over which it should be answered, each of the components of the triple is
lexicalized. The lexicalized subject and predicate of each triple are concatenated
together to form a textual query and use the lexicalized object as the answer.
(Note, this is trivally modi ed for the (?s, p, o case). We then retrieve the text
describing the subject. We assume that the text contains some reference to the
object under consideration.</p>
      <p>The location of that reference which we term an anchor is computed by the
given anchor function. For simplicity, in our implementation, we locate the rst
instance of the answer in the text. This may not always represent an instance
of the answer's lexical form which is located in an expression which answers
the speci c question form. More complex implementations could use di erent
heuristics or could return all possible anchor locations.</p>
      <p>We apply the algorithm to the combination of Wikipedia and Wikidata
dumps4. We attempted to obtain training data for all 1150 predicates in
Wikidata that associate two entities together. At this time, we do not treat predicates
that connect entities to literals. This is left for future work. We limited the
extraction to the top 20 entity type pairs per predicate, and limited each type pair
to 300 examples ). Thus, there is a maximum yield of 6000 examples per
predicate. We then apply the following cleaning/validation to the retrieved examples.
First, we drop examples where there is no Wikipedia page. Second, we ensure
that the answer is present in the Wikipedia page text. Finally, in order to ensure
adequate training data we lter out all models with less than 30 examples. Note
that this means that we have di ering amounts of training data per predicate.
After cleaning, we are able to obtain training data for 572 predicate for the
setting in which the object is the variable/answer. We term this the SP setting. On
average we have 929 examples per predicate with a maximum number of
examples of 5477 and a minimum of 30 examples. The median number of examples is
312. In the setting in which the subject is the variable / answer we are trying to
extract, enough data for 717 predicates is obtained. This is because the subject
answer is more likely to appear in the Wikipedia page text. We term this the
PO setting.</p>
    </sec>
    <sec id="sec-4">
      <title>Models</title>
      <p>
        Based on the above training data, we individual train models for all predicates
using the Jack the Reader framework [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. We use two state-of-the-art deep
learning architectures for extractive question answering, namely, FastQA [
        <xref ref-type="bibr" rid="ref28">28</xref>
        ] and the
implementation provided by the framework, JackQA. Both architectures are
interesting in that while they perform well on reading comprehension tasks (e.g.
SQuAD [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ]) both architectures try to eliminate complex additional layers and
thus have the potential for being modi ed in the future to suit this task. Instead
of describing the architectures in detail here, we refer the reader to
corresponding papers cited above. We also note that the Jack the Reader con guration les
provide succinct descriptions of the architectures, which are useful for
understanding their construction.
      </p>
      <p>To improve performance both in terms of reducing training time and to
reduce the amount of additional text the model training has to cope with, we
applied a windowing scheme. This is because longer text is normally associated
with greater issues when dealing with sequence models. Our scheme takes a
portion of the text around the answer location chosen from the Wikipedia content.
We now describe the following parameters for each architecture.</p>
      <p>
        FastQA All text is embedded using pre-trained GloVe word embeddings [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ]
(6 billion tokens, and 50 dimensions). We train for 10 epochs using a batch size
of 20. We constrain answers to be a maximum of 10 tokens and use a window
size of 1000 characters. The answer layer is con gured to be bilinear. We use the
ADAM optimizer with a learning rate of 0.11 and decay of 1.0.
      </p>
      <p>JackQA Here we embed the text using pre-trained GloVe word embeddings
(840 billion tokens and 300 dimensions). We use the default JackQA settings.
We use a window size of 3000 characters. The batch sizes were 128/96/64 for
three iterative runs. The subsequent runs with smaller batch sizes were only run
if the prior iteration failed. We speci ed a maximum number of 20 epochs.</p>
      <p>Baseline In addition to the models based on neural networks, we also
implemented a baseline. The baseline consisted of nding the closest noun phrase
to the property within the Wikipedia page and checking whether the answer is
contained within that noun phrase.</p>
      <p>Note, we attempted to nd functional settings that worked within our
available computational constraints. For example, FastQA requires more resources
than JackQA in relation to batch size , thus, we chose to use smaller
embeddings and window size in order to maintain a \good" batch size.</p>
      <p>We use 2/3 of the training data for model building and 1/3 for testing. Data
is divided randomly. Training was performed using an Amazon EC2 p2.xlarge5
box. It took 23 hours for training of FastQA models, which included all models
for all predicates even when there were too few training samples. For JackQA,
the window was increased to 3000 characters, and multiple training sessions were
required, reducing the batch size each time to complete the models which not
nish from earlier runs, in all three passes were required with 128, 96 and 64
batch size respectively. Total training time was 81 hours.</p>
      <p>Note that we train models for the setting where the subject and predicate
are bound but the object is not. We also use the FastQA architecture to build
models for the setting where the subject is treated as the variable to be bound.
5</p>
    </sec>
    <sec id="sec-5">
      <title>Experimental Results and Analysis</title>
      <p>Figure 3 plots individual model performance against the size of the training
data given. Overall, models based on deep learning notably outperform the
baseline models on average. Additionally, using these deep learning based approaches
we are able to create models that answer queries for 160 additional properties
over the baseline. In terms of analysis, rst, we wanted to see if there was a
correlation between the amount of training data and the performance of a model.
Using the data presented in Figure 3, we t a linear regression to it. We found no
statistically signi cant correlation (R2 = 0:37). The model architectures show
5000
s
lep4000
m
a
fxe3000
o
r
eb2000
m
uN1000</p>
      <p>0
0.0
0.2F1 Meas0u.4re (JackQ0A.6Models)0.8
1.0
0.0
0.2F1 Meas0u.r4e (FastQ0A.6Models)0.8
1.0
Models trained using Jack QA</p>
      <p>Models trained using Fast QA.
strong correlation in performance. The R2 value being 0:97 in the case of the F1
measure and 0:96 for the Exact measure. This suggests that the performance is
primarily a factor of the underlying kind of data. More details are provided in
Appendix A
6</p>
    </sec>
    <sec id="sec-6">
      <title>Prototype</title>
      <p>
        To understand whether this approach is feasible in practice, we implemented a
prototype of the system outlined in Figure 1. For the triple pattern fragment
facade we modify Piccolo, an open source triple pattern fragments server to
replace its in-memory based system with functions for calling out to our QA
answering component. The facade also implements a simple lexicalization
routine. The query answering component is implemented as a Python service and
calls out to an Elasticsearch search index where documents are stored. The query
answering component also pre-loads the models and runs each model across
candidate documents retrieved by querying elastic search. We also specify a max
number of candidate documents to run the models over. Currently, we execute
each model sequentially over all candidate documents. We then chose the top
set of ranked answers given the score produced by the model. Note that we can
return multiple bindings for the same ranked results. We made some
preliminary timing estimates of a query. It takes on the order of 10 seconds to provide
results for a single triple pattern query. This is surprisingly good given the fact
that we execute models sequentially instead of in parallel. Furthermore, we
execute the models over the entirety of the Wikipedia article. Our own anecdotal
experience shows that question answering models are both faster and produce
more accurate results when supplied with smaller amounts of text. Thus, there is
signi cant room for optimizing query performance with some simple approaches
including parallelizing models, chunking text into smaller blocks, and limiting
the number of models executed to those that are speci c for the triple pattern.
Furthermore, it is straightforward to issue triple pattern fragment query requests
over multiple running instances [
        <xref ref-type="bibr" rid="ref25">25</xref>
        ]. One could also implement more complex
sharding mechanisms designed for triple tables [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. Overall, the prototype gives
us con dence that this sort of system could be implemented practically.6
7
      </p>
    </sec>
    <sec id="sec-7">
      <title>Related Work</title>
      <p>
        Our work builds upon and connects to a number of existing bodies of
literature. The work on information extraction is closely related. [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ] provides a
recent survey of the literature in this area speci cally targeted to the the
problems of extracting and linking of entities, concepts and relations. One can view
the models that we build as similar to distantly supervised relation extraction
approaches [
        <xref ref-type="bibr" rid="ref16 ref23">16, 23</xref>
        ], where two mentions of entities are found in text and the
context around those mentions is used to learn evidence for that relation. Recent
approaches have extended the notion of context [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ] and applied neural networks
to extract relations [
        <xref ref-type="bibr" rid="ref30 ref6">30, 6</xref>
        ].
      </p>
      <p>
        The closest work to ours in the information extraction space is [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] where they
apply machine comprehension techniques to extract relations. Speci cally, they
translate relations into templated questions - a process they term queri cation.
For example, for the relation spouse(x,y) they created a series of
corresponding question templates such as \Who is x married to?". These templates are
constructed using crowdsourcing, where the workers are provided a relation,
example sentence and asked to produce a question template. This dataset is used
to train a BiDAF-based model [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ] and similar to our approach they address slot
lling queries where the aim is to populate one side of the relation. While we
apply a similar technique, our approach di ers in a two key aspects. First, we
target a di erent task, namely, answering structured queries. Second, we do not
generate questions through question templates but instead build the questions
out of the knowledge base itself.
      </p>
      <p>
        Like much of the work in this space our approach is based on a large scale
parallel corpus. Of particular relevance to our task are the WikiSQL and
WikiReading corpora. WikiSQL [
        <xref ref-type="bibr" rid="ref31">31</xref>
        ] provides a parallel corpus that binds SQL queries to
a natural language representation. The task the dataset is used for is to
answer natural language questions over SQL unlike ours which is to answer
SQLlike queries over text. SQLWikiReading [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] like our approach extracts a corpus
from Wikidata and Wikipedia in order to predict the value of particular
properties. Another corpus of note is ComplexWebQuestions [
        <xref ref-type="bibr" rid="ref24">24</xref>
        ], which pairs complex
SPARQL queries with natural language queries. Importantly, it looks at the
compositionality of queries from smaller units. Like WikiSQL, it looks at answering
natural language queries over databases. In general, we think our approach in
also specifying an extraction procedure is a helpful addition for applying corpus
construction in di erent domains.
      </p>
      <p>As mentioned in the introduction, text databases, where information
extraction is combined with databases are also relevant. Our system architecture was</p>
      <sec id="sec-7-1">
        <title>6 We also integrated the prototype with Slack.</title>
        <p>
          inspired by the pioneering work of [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ]. In that work, a search index is used to
rst locate potential documents and then information extraction techniques are
applied to the selected documents to populate a database. Our approach di ers
in two key aspects. First, instead of populating a database our system substitutes
the indexes of the database with models. Second, we use distributed query
techniques in order to process complex queries on the client side. Recent work [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ]
uses deep learning based approaches to perform information extraction during
database query execution speci cally for entity disambiguation. Similar to other
work in this area, and unlike ours, they integrate the information extraction
within the database engine itself.
        </p>
        <p>
          Finally, there is a long history of mixing information retrieval and database
style queries together. For example, for the purposes of querying over
semistructured data [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ]. [
          <xref ref-type="bibr" rid="ref19">19</xref>
          ] provides an accessible introduction to that history. While our
system is designed to answer database queries one can imagine easily extending
to the semistructured setting.
8
        </p>
      </sec>
    </sec>
    <sec id="sec-8">
      <title>Conclusion &amp; Future Work</title>
      <p>In this work, we have explored the notion of answering database queries over
text absent the need for a traditional database intermediary. We have shown
that this approach is feasible in practice by combining machine comprehension
based models with distributed query techniques.</p>
      <p>
        There are a number of avenues for future work. In the short term, the
developed models could be expanded to include extracting properties as well as
subjects and objects. We also think that joint models for all triple pattern
predictions is worth exploring. One would also want to extend the supported queries
to consider not only relationships between entities but also to the attributes of
entities. Our current lexicalization approach is also quite simple and could be
improved by considering it as the inverse of the entity linking problem and applying
those techniques or applying summarization approaches [
        <xref ref-type="bibr" rid="ref26">26</xref>
        ]. In this work, we
used model architectures that are designed for answering verbalized questions
and not database queries. Modifying these architectures may also be a
direction to obtain even better performance. Obviously more extensive experimental
evaluations would be of interest, in particular, extending the approach to other
knowledge bases and looking more deeply at query result quality.
      </p>
      <p>
        In the long term, the ability to query over all types of data whether images,
structured data or text has proven useful for knowledge bases [
        <xref ref-type="bibr" rid="ref29">29</xref>
        ]. Extending
our concept to deal with these other datatypes could be powerful -making it
easy to perform structured queries over unstructured data while minimizing
information extraction overhead. In general, we believe that structured queries
will continue to be a useful mechanism for data professionals to both work with
data and integrate information into existing data pipelines. Hence, focusing on
automated knowledge base construction from the query vantage point is an
important perspective.
A
      </p>
    </sec>
    <sec id="sec-9">
      <title>Individual Model and Error Analysis</title>
      <p>We looked more deeply at performance for individual models for a given property.
Table 3 shows the highest performing models. We nd some consistent patterns.
First, properties that have speci c value constraints within Wikidata generate
good results. For example, the \crystal system" property needs to have one of 10
values (e.g cubic crystal system, quasicrystal, amorphous solid). Likewise, the
\coolant" property needs to be assigned one of fourteen di erent values (e.g.
water, oil, air). This is also true of \discovery method", which oddly enough is
actually de ned as the the method by which an exoplanet is discovered. This is
also a feature of properties whose values come from classi cation systems (e.g.
\Kppen climate classi cation" and "military casualty classi cation").</p>
      <p>A second feature that seems to generate high performing models are those
that refer to common simple words. For example, the \source of energy" property
takes values such as \wind" or \human energy".</p>
      <p>Lastly, simple syntactic patterns seem to be learned well. For example, the
property "birthday", which links to entities describing a month, day combination
(e.g. November 8) which is thus restricted to a something that looks like a month
string followed by one or two numerical characters. Likewise, the expected value
for the property \ ag" often appears directly in text itself. That is the correct
answer for the query \Japan ag" is \ ag of Japan", which will appear directly
in text.</p>
      <p>We also look at the lowest performing models, shown in Table 4 to see what is
di cult to learn. Ratings for lms (e.g. Australian Classi cation, RTC lm
rating, EIRIN lm rating) seem extremely di cult to learn. Each of these properties
expect values of two or three letters (e.g. PG, R15+, M). The property \blood
type" also has the same form. It seem that using character level embeddings
may worked better in these cases.</p>
      <p>The property \contains administrative territorial entity " is an interesting
case as there are numerous examples. This property is used within Wikidata to
express the containment relation in geography. For example, that county contains
a village or a country contains a city. We conjecture that this might be di cult
to learn because the sheer variety of linkages that this can express making it
di cult to nd consistencies in the space. A similar issue could be present for
properties such as \voice actor" and \cast member" where the values can be
essentially any person entity. Similarly, \polymer of" and \species kept" both
can take values that come from very large sets (e.g. all chemical compounds
and all species). It might be useful for the model to be provided speci c hints
about types (i.e. actors, chemicals, locations) that may allow it to nd indicative
features.</p>
      <p>Property</p>
      <p>Fast QA Fast QA Jack QA Jack QA Training</p>
      <p>F1 Exact F1 Exact Data Size
0.00 0.00 0.00 0.00 48
0.00 0.00 0.00 0.00 167
0.01 0.01 0.02 0.02 349
0.00 0.00 0.08 0.08 302
0.09 0.06 0.08 0.07 1838</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Abdelaziz</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Harbi</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Khayyat</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kalnis</surname>
            ,
            <given-names>P.:</given-names>
          </string-name>
          <article-title>A survey and experimental comparison of distributed sparql engines for very large rdf data</article-title>
          .
          <source>Proceedings of the VLDB Endowment</source>
          <volume>10</volume>
          (
          <issue>13</issue>
          ),
          <year>2049</year>
          {
          <year>2060</year>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Abiteboul</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Querying semi-structured data</article-title>
          .
          <source>In: International Conference on Database Theory</source>
          . pp.
          <volume>1</volume>
          {
          <fpage>18</fpage>
          . Springer (
          <year>1997</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Alagiannis</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Borovica</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Branco</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Idreos</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ailamaki</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Nodb: e cient query execution on raw data les</article-title>
          .
          <source>In: Proceedings of the 2012 ACM SIGMOD International Conference on Management of Data</source>
          . pp.
          <volume>241</volume>
          {
          <fpage>252</fpage>
          .
          <string-name>
            <surname>ACM</surname>
          </string-name>
          (
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Cafarella</surname>
            ,
            <given-names>M.J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Re</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Suciu</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Etzioni</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Banko</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Structured querying of web text</article-title>
          .
          <source>In: 3rd Biennial Conference on Innovative Data Systems Research (CIDR)</source>
          , Asilomar, California, USA (
          <year>2007</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <given-names>Dirk</given-names>
            <surname>Weissenborn</surname>
          </string-name>
          , Pasquale Minervini,
          <string-name>
            <given-names>T.D.I.A</given-names>
            .
            <surname>J.W.T.R.M.B.J.M.T.D.P.S.S.R.:</surname>
          </string-name>
          <article-title>Jack the Reader A Machine Reading Framework</article-title>
          . In:
          <article-title>Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (ACL) System Demonstrations</article-title>
          (
          <year>July 2018</year>
          ), https://arxiv.org/abs/
          <year>1806</year>
          .08727
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Glass</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gliozzo</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hassanzadeh</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mihindukulasooriya</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rossiello</surname>
          </string-name>
          , G.:
          <article-title>Inducing implicit relations from text using distantly supervised deep nets</article-title>
          .
          <source>In: International Semantic Web Conference</source>
          . pp.
          <volume>38</volume>
          {
          <fpage>55</fpage>
          . Springer (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Harris</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Seaborne</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Prudhommeaux</surname>
          </string-name>
          , E.:
          <article-title>Sparql 1.1 query language</article-title>
          .
          <source>W3C recommendation 21(10)</source>
          (
          <year>2013</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Hewlett</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lacoste</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jones</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Polosukhin</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fandrianto</surname>
            , A., Han,
            <given-names>J</given-names>
          </string-name>
          .,
          <string-name>
            <surname>Kelcey</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Berthelot</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>WIKIREADING: A novel large-scale language understanding task over Wikipedia</article-title>
          .
          <source>In: Proceedings of the The 54th Annual Meeting of the Association for Computational Linguistics (ACL</source>
          <year>2016</year>
          )
          <article-title>(</article-title>
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Ho</surname>
            <given-names>mann</given-names>
          </string-name>
          , R.,
          <string-name>
            <surname>Zettlemoyer</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Weld</surname>
            ,
            <given-names>D.S.:</given-names>
          </string-name>
          <article-title>Extreme extraction: Only one hour per relation</article-title>
          .
          <source>arXiv preprint arXiv:1506.06418</source>
          (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Jain</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Doan</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gravano</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          :
          <article-title>Sql queries over unstructured text databases</article-title>
          .
          <source>In: Data Engineering</source>
          ,
          <year>2007</year>
          .
          <article-title>ICDE 2007</article-title>
          . IEEE 23rd International Conference on. pp.
          <volume>1255</volume>
          {
          <fpage>1257</fpage>
          .
          <string-name>
            <surname>IEEE</surname>
          </string-name>
          (
          <year>2007</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Kilias</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          , Loser,
          <string-name>
            <given-names>A.</given-names>
            ,
            <surname>Andritsos</surname>
          </string-name>
          ,
          <string-name>
            <surname>P.</surname>
          </string-name>
          :
          <article-title>Indrex: In-database relation extraction</article-title>
          .
          <source>Information Systems</source>
          <volume>53</volume>
          ,
          <fpage>124</fpage>
          {
          <fpage>144</fpage>
          (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Kilias</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          , Loser,
          <string-name>
            <given-names>A.</given-names>
            ,
            <surname>Gers</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.A.</given-names>
            ,
            <surname>Koopmanschap</surname>
          </string-name>
          ,
          <string-name>
            <surname>R.</surname>
          </string-name>
          , Zhang,
          <string-name>
            <given-names>Y.</given-names>
            ,
            <surname>Kersten</surname>
          </string-name>
          ,
          <string-name>
            <surname>M.</surname>
          </string-name>
          :
          <article-title>Idel: In-database entity linking with neural embeddings</article-title>
          . arXiv preprint arXiv:
          <year>1803</year>
          .
          <volume>04884</volume>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Kumar</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Irsoy</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ondruska</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Iyyer</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bradbury</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gulrajani</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhong</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Paulus</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Socher</surname>
          </string-name>
          , R.:
          <article-title>Ask me anything: Dynamic memory networks for natural language processing</article-title>
          .
          <source>In: International Conference on Machine Learning</source>
          . pp.
          <volume>1378</volume>
          {
          <issue>1387</issue>
          (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Levy</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Seo</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Choi</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zettlemoyer</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          :
          <article-title>Zero-shot relation extraction via reading comprehension</article-title>
          .
          <source>In: Proceedings of the 21st Conference on Computational Natural Language Learning (CoNLL</source>
          <year>2017</year>
          ). pp.
          <volume>333</volume>
          {
          <issue>342</issue>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Martinez-Rodriguez</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hogan</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lopez-Arevalo</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          :
          <article-title>Information extraction meets the semantic web: A survey</article-title>
          .
          <source>Semantic Web Journal</source>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <surname>Mintz</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bills</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Snow</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jurafsky</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>Distant supervision for relation extraction without labeled data</article-title>
          .
          <source>In: Proceedings of the Joint Conference of the 47th Annual Meeting of the ACL and the 4th International Joint Conference on Natural Language Processing of the AFNLP: Volume 2-</source>
          Volume 2. pp.
          <volume>1003</volume>
          {
          <fpage>1011</fpage>
          .
          <article-title>Association for Computational Linguistics (</article-title>
          <year>2009</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          17.
          <string-name>
            <surname>Pennington</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Socher</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Manning</surname>
          </string-name>
          , C.D.: Glove:
          <article-title>Global vectors for word representation</article-title>
          .
          <source>In: Empirical Methods in Natural Language Processing (EMNLP)</source>
          . pp.
          <volume>1532</volume>
          {
          <issue>1543</issue>
          (
          <year>2014</year>
          ), http://www.aclweb.org/anthology/D14-1162
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          18.
          <string-name>
            <surname>Quirk</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Poon</surname>
          </string-name>
          , H.:
          <article-title>Distant supervision for relation extraction beyond the sentence boundary</article-title>
          .
          <source>In: Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics: Volume</source>
          <volume>1</volume>
          ,
          <string-name>
            <given-names>Long</given-names>
            <surname>Papers</surname>
          </string-name>
          . vol.
          <volume>1</volume>
          , pp.
          <volume>1171</volume>
          {
          <issue>1182</issue>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          19.
          <string-name>
            <surname>Raghavan</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Garcia-Molina</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          :
          <article-title>Integrating diverse information management systems: A brief survey</article-title>
          .
          <source>Bulletin of the Technical Committee on Data Engineering</source>
          p.
          <volume>44</volume>
          (
          <year>2001</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          20.
          <string-name>
            <surname>Rajpurkar</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          , Zhang, J.,
          <string-name>
            <surname>Lopyrev</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Liang</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          : Squad:
          <volume>100</volume>
          ,000+
          <article-title>questions for machine comprehension of text</article-title>
          .
          <source>In: Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing</source>
          . pp.
          <volume>2383</volume>
          {
          <issue>2392</issue>
          (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          21.
          <string-name>
            <surname>Seo</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kembhavi</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Farhadi</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hajishirzi</surname>
          </string-name>
          , H.:
          <article-title>Bidirectional attention ow for machine comprehension</article-title>
          .
          <source>arXiv preprint arXiv:1611.01603</source>
          (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          22.
          <string-name>
            <surname>Shin</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wu</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>De Sa</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhang</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Re</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>Incremental knowledge base construction using deepdive</article-title>
          .
          <source>Proceedings of the VLDB Endowment</source>
          <volume>8</volume>
          (
          <issue>11</issue>
          ),
          <volume>1310</volume>
          {
          <fpage>1321</fpage>
          (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          23.
          <string-name>
            <surname>Surdeanu</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tibshirani</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nallapati</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Manning</surname>
          </string-name>
          , C.D.
          <article-title>: Multi-instance multilabel learning for relation extraction</article-title>
          .
          <source>In: Proceedings of the</source>
          <year>2012</year>
          <article-title>joint conference on empirical methods in natural language processing and computational natural language learning</article-title>
          . pp.
          <volume>455</volume>
          {
          <fpage>465</fpage>
          .
          <article-title>Association for Computational Linguistics (</article-title>
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          24.
          <string-name>
            <surname>Talmor</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Berant</surname>
            ,
            <given-names>J.:</given-names>
          </string-name>
          <article-title>The web as a knowledge-base for answering complex questions</article-title>
          .
          <source>In: Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</source>
          , Volume
          <volume>1</volume>
          (
          <string-name>
            <given-names>Long</given-names>
            <surname>Papers</surname>
          </string-name>
          <article-title>)</article-title>
          .
          <source>vol. 1</source>
          , pp.
          <volume>641</volume>
          {
          <issue>651</issue>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          25.
          <string-name>
            <surname>Verborgh</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sande</surname>
            ,
            <given-names>M.V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hartig</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Herwegen</surname>
            ,
            <given-names>J.V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vocht</surname>
            ,
            <given-names>L.D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Meester</surname>
            ,
            <given-names>B.D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Haesendonck</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Colpaert</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Triple pattern fragments: A low-cost knowledge graph interface for the web</article-title>
          .
          <source>Web Semantics: Science, Services and Agents on the World Wide Web 37-38</source>
          ,
          <issue>184</issue>
          {
          <fpage>206</fpage>
          (
          <year>2016</year>
          ). https://doi.org/https://doi.org/10.1016/j.websem.
          <year>2016</year>
          .
          <volume>03</volume>
          .003, http://www.sciencedirect.com/science/article/pii/S1570826816000214
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          26.
          <string-name>
            <surname>Vougiouklis</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Elsahar</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ka</surname>
            <given-names>ee</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>L.A.</given-names>
            ,
            <surname>Gravier</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            ,
            <surname>Laforest</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            ,
            <surname>Hare</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            ,
            <surname>Simperl</surname>
          </string-name>
          ,
          <string-name>
            <surname>E.</surname>
          </string-name>
          :
          <article-title>Neural wikipedian: Generating textual summaries from knowledge base triples</article-title>
          .
          <source>Journal of Web Semantics</source>
          (
          <year>2018</year>
          ). https://doi.org/https://doi.org/10.1016/j.websem.
          <year>2018</year>
          .
          <volume>07</volume>
          .002, http://www.sciencedirect.com/science/article/pii/S1570826818300313
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          27.
          <string-name>
            <surname>Vrandecic</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>Wikidata: A new platform for collaborative data collection</article-title>
          .
          <source>In: Proceedings of the 21st International Conference on World Wide Web</source>
          . pp.
          <volume>1063</volume>
          {
          <fpage>1064</fpage>
          .
          <string-name>
            <surname>ACM</surname>
          </string-name>
          (
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          28.
          <string-name>
            <surname>Weissenborn</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wiese</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          , Sei e, L.:
          <article-title>Making neural qa as simple as possible but not simpler</article-title>
          .
          <source>In: Proceedings of the 21st Conference on Computational Natural Language Learning (CoNLL</source>
          <year>2017</year>
          ). pp.
          <volume>271</volume>
          {
          <issue>280</issue>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          29.
          <string-name>
            <surname>Wu</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hsiao</surname>
            , L., Cheng,
            <given-names>X.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hancock</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rekatsinas</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Levis</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Re</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>Fonduer: Knowledge base construction from richly formatted data</article-title>
          .
          <source>In: Proceedings of the 2018 International Conference on Management of Data</source>
          . pp.
          <volume>1301</volume>
          {
          <fpage>1316</fpage>
          .
          <string-name>
            <surname>ACM</surname>
          </string-name>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          30.
          <string-name>
            <surname>Zeng</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Liu</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lai</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhou</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhao</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          :
          <article-title>Relation classi cation via convolutional deep neural network</article-title>
          .
          <source>In: Proceedings of COLING</source>
          <year>2014</year>
          ,
          <source>the 25th International Conference on Computational Linguistics: Technical Papers</source>
          . pp.
          <volume>2335</volume>
          {
          <issue>2344</issue>
          (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          31.
          <string-name>
            <surname>Zhong</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Xiong</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Socher</surname>
          </string-name>
          , R.: Seq2sql:
          <article-title>Generating structured queries from natural language using reinforcement learning</article-title>
          .
          <source>CoRR abs/1709</source>
          .00103 (
          <issue>2017</issue>
          )
          <article-title>Australian Classi cation RTC lm rating EIRIN lm rating blood type contains administrative territorial entity</article-title>
          <source>voice actor 0.12 0.11 0.11 0.09 2562 species kept 0.11 0.03 0.12 0.03 354 best sprinter classi cation 0.19 0.18 0.12 0.11 165 cast member 0.15 0.14 0.13 0.11 3955 polymer of 0.18 0.08 0.13 0.08 38 Table</source>
          <volume>4</volume>
          :
          <article-title>Lowest 10 performing models in the SP setting as determined by F1 measures from models trained using the Jack QA architecture</article-title>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>