=Paper=
{{Paper
|id=Vol-1180/CLEF2014wn-QA-ShizhuEt2014
|storemode=property
|title=CASIA@V2: A MLN-based Question Answering System over Linked Data
|pdfUrl=https://ceur-ws.org/Vol-1180/CLEF2014wn-QA-ShizhuEt2014.pdf
|volume=Vol-1180
|dblpUrl=https://dblp.org/rec/conf/clef/HeZL014
}}
==CASIA@V2: A MLN-based Question Answering System over Linked Data==
CASIA@V2: A MLN-based Question Answering
System over Linked Data
Shizhu He, Yuanzhe Zhang, Kang Liu, and Jun Zhao
National Laboratory of Pattern Recognition,
Institute of Automation, Chinese Academy of Sciences.
Zhongguancun East Road 95#, Beijing, 100084, China.
{shizhu.he,yzzhang,kliu,jzhao}@nlpr.ia.ac.cn
Abstract. We present a question answering system (CASIA@V2) over Linked
Data (DBpedia), which translates natural language questions into structured queries
automatically. Existing systems usually adopt a pipeline framework, which con-
tains four major steps: 1) Decomposing the question and detecting candidate
phrases; 2) mapping the detected phrases into semantic items of Linked Data;
3) grouping the mapped semantic items into semantic triples; and 4) generat-
ing the rightful SPARQL query. We present a jointly learning framework using
Markov Logic Network(MLN) for phrase detection, phrases mapping to seman-
tic items and semantic items grouping. We formulate the knowledge for resolving
the ambiguities in three steps of QALD as first-order logic clauses in a MLN. We
evaluate our approach on QALD-4 test dataset and achieve an F-measure score of
0.36, an average precision of 0.32 and an average recall of 0.40 over 50 questions.
Keywords: Question Answering, Linked Data, Markov Logic Network
1 Introduction
With the rapid development of the Web of Data, there are many RDF datasets pub-
lished as Linked Data [1], such as DBpedia [2], Freebase [3] and YAGO [4]. The grow-
ing Linked Data contain a wealth of knowledge including entities, classes and proper-
ties. Moreover, these linked data usually have complex structures and are highly het-
erogeneous. However, there are the gaps between the users and the Linked Data. On the
one hand, the users, even expert programmers, need a lot of practices to handle stan-
dard structured query languages like SPARQL. On the other hand, due to the diversity
In: Cappellato, L., Ferro, N., Halvey, M., and Kraaij, W., editors (2014). CLEF 2014 Labs and
Workshops, Notebook Papers. CEUR Workshop Proceedings (CEUR-WS.org), ISSN 1613-
0073, http://ceur-ws.org/ Vol-1180/.
Thanks to Prof. Hang Li, Dr. Yibo Zhang and Dr. Jie Zhang at Huawei Noah’S Ark Lab and the
anonymous reviewers for their insightful advices and and suggestions. This work was spon-
sored by the National Natural Science Foundation of China (No. 61272332 and No. 61202329)
and CCF-Tencent Open Research Fund. This work was supported in part by Noah’s Ark Lab
of Huawei Tech. Ltm.
1249
and high heterogeneity of the Linked Data, it is difficult for humans to select relevant
resources and discover useful information. Thus, developing user-friendly interface for
accessing those linked data become increasing important.
Question answering over Linked Data [5] is aimed at eliminating those gaps, which
attempts to allow the users to access those structured data with natural language. For
example, with respect to the question: “Which software has been developed by orga-
nizations founded in California, USA?”, the aim is to automatically convert this ut-
terance into a SPARQL query which contains the following subject-property-object
(SPO) triple format: h?url rdf:type dbo:Software, ?url dbo:developer ?x1, ?x1 rdf:type
dbo:Company, ?x1 dbo:foundationPlace dbr:Californiai1 .
Lopez et al. [6] have given a comprehensive survey in this research area. The authors
develop PowerAqua system [7] to answer questions on large, heterogeneous datasets.
For questions containing quantifiers, comparatives or superlatives, Unger et al. [8] trans-
late natural language(NL) questions to formal language(FL) structured query using sev-
eral SPARQL templates and a set of heuristic rules mapping phrase to semantic items.
And DEANNA [9] jointly disambiguates the following tasks based on integer linear pro-
gramming: segmenting question, mapping phrases to semantic items and constructing
SPARQL triple patterns.
This paper proposes a novel algorithm based on a learning framework, Markov
Logic Network (MLN) [10], to learn a joint model for constructing structured queries
from natural language utterances. MLN is a statistical relational learning framework
that combines first-order logic and Markov networks. The appealing property of MLN
is that it is readily interpretable by humans and is natural to perform joint learning under
the Markov logic framework. We formulate the knowledge for resolving the ambigu-
ities in three steps of QALD (phrase detection, phrase-to-semantic-item mapping and
semantic items grouping) as first-order logic clauses in a MLN. In the framework of
MLN, all clauses will make interacted effects which combines resolving all problems
into a unified process. In this way, the result in each step can be globally optimized.
Moreover, different from previous methods, we adopt a learning strategy to automat-
ically learn the patterns for semantic items grouping. We also formulate the learned
patterns as first-order clauses in MLN. The model will learn the weights of each clause
to determine the most effective patterns for semantic triples construction. In this way,
our approach can cover more semantic expressions and answer more questions than
previous method based on manually designed patterns.
We evaluate our approach on QALD-4 test dataset and achieve an F-measure score
of 0.36, an average precision of 0.32 and an average recall of 0.40 over 50 questions.
2 System Description
The current version of our QA system is not designed to answer the questions which
contain numbers, date comparisons and aggregation operations such as group by or
order by. We also do not consider the questions which contain filter condition. Figure
1
prefixes in semantic items indicate the source of its vocabularies, dbr indi-
cate entity and dbo indicate class and property defined in the DBpedia ontology
(http://wiki.dbpedia.org/Ontology39).
1250
Fig. 1. The architecture of CASIA@V2
1 shows the architecture of our system to translate a question into a formal SPARQL
query.
At first, sequence of tokens(phrase) are detected that probably indicate semantic
items, such as software, developed by and California. This step detects full potentially
phrases, and leaves the decision for phrases selecting in later steps.
Next, the phrases are mapping to semantic items. Phrases can denote entities, classes
and properties. For example, the phrase software can denote class dbo:Software and
property dbo:developer, and the phrase developed by can denote entity
dbr:videogamedeveloper, class dbo:BritishRoyalty and property dbo:developer. This step
purely constructs a candidate space for every possible mapping, and leaves the decision
for select which semantic items in the next step.
Then, we should make the decisions for choosing phrases, mapping the chosen
phrases to suitable semantic items and determine the relations of selected semantic
item. We formulate the joint decisions as an generalized inference task. We employ rich
features and constraints (including hard and soft constraints) to infer the joint decisions
using a MLN.
Finally, with the inference results, the last step constructs a semantic item query
graph, and generates an executable SPARQL query with the question type.
We will give a detailed description of each component and give a step by step ex-
planation with the following example, Figure 2 shows the intermediate results of every
steps:
Which software has been developed by organizations founded in California, USA?.
1251
Which software has been developed by
organizations founded in California, USA? software, developed, developed by, organizations,
founded, founded in, California, USA
software California
phraseIndex priorMatchScore
phraseDepTag
developed by
... resourceType
...
...
hasRelatedness
isTypeCompatible
phrasePosTag
hasPhrase hasResource
hasMeanWord
hasRelation
...
Fig. 2. Framework of our system.
2.1 System Pipleline
1) Phrase detection. Sequences of tokens (phrases) that probably indicate semantic
items are detected. To this end, we do not use named entity recognizer (NER) because of
its low coverage. 2 To avoid missing useful phrases, we retain all n-grams as candidate
phrases, and then use some rules to filter them. The rules include: the length of tokens
span must be less than 4 (excepting all contiguous tokens are capitalizations); the POS
tag of the start token must be jj, nn, rb and vb; all contiguous capitalization tokens must
not be split, and so on. For instance, software, developed by, organizations, founded in
and California are detected in the example.
2) Mapping phrase to semantic item. After phrases are detected, each phrase may
be mapped to the semantic items in KB (entities, classes and properties). For exam-
ple, software is mapped to dbo:Software, dbo:developer, etc.; California is mapped to
2
We perform testing in two commonly used question corpus (QALD-3 Training data
and free917) using Stanford CRF-based NER tool(http://nlp.stanford.edu/software/CRF-
NER.shtml). The results demonstrate that merely 51.5% and 23.8% right NEs can be rec-
ognized, respectively.
1252
dbr:California, dbr:California (wine), etc. We use different techniques and resources
to map phrases to different types of semantic items. For mapping phrases to entities,
considering the entities in DBpeida are curated from Wikipeida, we employ anchors,
redirections and disambiguations information from Wikipedia. For mapping phrases
to classes, considering that classes have lexical variation, especially synonyms, e.g.,
dbo:Film could be mapped from film, movie and show, we use word2vec tool3 to convert
every word (phrase) into a vector and compute the similarity between the phrase and
the class in KB. The similarity scoring methods are introduced in Section 3.2. Then,
the top-N most similar classes for each phrase are returned. For mapping phrases to
properties, we employ the resources from PATTY [11] and ReVerb [12]. Specifically,
we first compute the associations between the semantic properties in DBpedia and rela-
tion patterns in PATTY and ReVerb through instance alignments as same as [13]. Next,
if a detected phrase is matched to some relation pattern, the corresponding properties
in DBpedia will be returned as the candidates. This step purely constructs a candidate
space for every possible mapping, and the decision of selecting the best fitting semantic
item is made in the next step.
3) Feature extraction and joint inference. There are ambiguities in phrase detec-
tion and mapping-phrase-to-semantic-item. This step consists in the resolution of these
ambiguities and determine the relations among the mapped semantic items. It is the
core contribution of this paper, which performs disambiguation in a unified manner.
First, feature extraction is performed to prepare rich features from the question and the
knowledge base. Next, the disambiguation is performed in a joint fashion with a Markov
Logic Network (MLN). The detailed information will be presented in the next Section.
4) SPARQL generation. Based on the inference results, we construct a query
graph. The vertex contains the following information: the phrase, token span indexs
of the phrase, the mapped semantic item and its type. The edge indicates the relation
between two semantic items. For example, we use 1 2 to indicate that the first argu-
ment of an item matches the second argument of another item4 . The right bottom in
Figure 2 shows an example of it. The relation in the query graph is paired data merely,
but the SPARQL queries need the grouped triples of semantic items. Thus, we con-
vert a query graph into multiple joined semantic triples. Three interconnected semantic
items, which must ensure the middle item is a property, are converted into a seman-
tic triple. For example, the query graph Vdbo:Book[Class] 1← →
2 dbo:author[Property] 1 1
←→
dbr:Danielle Steel[Entity]W is converted into h?x rdf:type dbo:Book, dbr:Danielle dbo:author
?xi, and Vdbo:populationTotal[Property] 1 2 dbo:capital[Property] 1 1 dbr:Australia[Entity]W5
←→ ←→
is converted into h?x1 dbo:populationTotal ?answer, ?x1 dbo:capital dbr:Australiai. If the
query graph only contains one vertex which indicates a class ClassURI, we generate h
?x rdf:type ClassURIi. If the query graph contains two connected vertexes, we append
a variable to bind the missing match argument of the semantic item.
The final SPARQL query is constructed by joining the semantic item triples and
combining them with the corresponding SPARQL template. We divide the questions
into three types: Yes/No, Number and Normal. Yes/No questions use the ASK WHERE
3
https://code.google.com/p/word2vec/
4
The other marks will be introduced in Section 3.1.
5
corresponding the question “How many people live in the capital of Australia?”
1253
template. With respect to number questions, we use SELECT COUNT(?url) WHERE
template, if it cannot obtain a fitting SPARQL query(the query result are not a number),
we then use normal question template to generate a query again. Normal questions
use the SELECT ?url WHERE template. For instance, we construct the SPARQL query
SELECT(?url) WHERE{ ?url rdf:type dbo:Software. ?url dbo:developer ?x1. ?x1 rdf:type
dbo:Company. ?x1 dbo:foundationPlace dbr:California.} toward the example.
3 Feature extraction & joint inference
In this section, we will first briefly describe Markov Logic Networks. Then, we
present the predicates(features) and the first-order logic formulas for joint inference.
3.1 Markov Logic Networks
Markov logic networks combine Markov networks with first-order logic in proba-
bilistic framework[10,14]. A Markov logic network consists of a set of first-order for-
mulas. Each formula consists of a set of first-order predicates, logical connectors and
variables. However, differently to first-order logic where a formula represents a hard
constraints, these logic formulas are softened and can be violated with some penalty
(the weight of formula) in MLN.
An MLN M is consists of several weighted formulas {(φi , wi )}i , where φi is a first
order formula and wi is the penalty(the formula’s weight). These weighted formulas
define a probability distribution over sets of possible worlds. Let y denote a possible
world, then p(y) is defined as follows:
1 X X
p(y) = exp wi fcφi (y) , (1)
Z nφ
(φi ,wi )∈M c∈C i
where each c is a binding of free variables in φi to constants; fcφi is a binary feature
function that returns 1 if the ground formula we get through replacing the free variables
in φi with the constants in c under the given possible world y is true, and 0 otherwise;
C nφi is the set of all possible bindings for the free variables in φi . Z is a normalization
constant. The Markov Network corresponds to this distribution, where nodes represent
ground atoms and factors represent ground formulas.
Let us illustrate how MLN determine the relation of semantic items mapped by
phrases. The following formulas indicates that if two semantic items have some depen-
dency path tags6 , then they have some type of relations with some weights.
(φ1 , w1 ) : depP athT ag(a, b, “pobj”) ∧ (a 6= b)
(2)
⇒ relation(a, b, “2 1”).
(φ2 , w2 ) : depP athT ag(a, b, “pobj”) ∧ (a 6= b)
(3)
⇒ relation(a, b, “1 1”).
6
Actually, the dependency path tags are extracted from the phrases which mapped to semantic
items. Here is simplified to illustrate how MLN works.
1254
Here, a and b are variables which represent any semantic item, depPathTag and relation
are an observed predicate and a hidden predicate, respectively. The values of observed
predicates are known from feature extraction, and the values of hidden predicates are
infered. The values of two weights w1 and w2 affect the decision of choosing relation
types between two semantic items.
There are a lot of methods to inference and learn the weights for MLN[14,10].
Several packages for MLN learning available online for free, such as Alchemy7 , Tuffy8 ,
thebeast9 .
3.2 Predicates
In MLN, we design several predicates to resolve the ambiguities in phrase detec-
tion, mapping phrases to semantic items and semantic items grouping. Specifically, we
design a hidden predicate hasPhrase(i) to indicate that the ith candidate phrase has been
chosen, predicate hasResouce(i,j) to indicate that the ith phrase is mapped to the jth se-
mantic item, and predicate hasRelation(j,k,rr) to indicate that the jth semantic item and
the kth semantic item can be grouped together with the relation type rr. Note that we
define four relation types between two semantic items: 1 1, 1 2, 2 1 and 2 2. Here, the
relation type t s means the tth argument of the first semantic item corresponds to the sth
argument of the second semantic item. The detailed illustration is shown in Table 1.
Table 1. Examples of the relation types.
Type Example Question
11 dbo:height 1 1 dbr:Michael Jordan How tall is Michael Jordan?
12 dbo:River 1 2 dbo:crosses Which river does the Brook-
lyn Bridge cross?
21 dbo:creator 2 1 dbr:Walt Disney Which spaceflights were
launched from Baikonur?
22 dbo:birthPlace 2 2 dbo:capital Which actors were born in
the capital of American?
Moreover, we define a set of observed predicates to describe the properties of phrases,
semantic items, relations between phrases and relations between semantic items. The
observed predicates and descriptions are shown in Table 2.
Previous methods usually designed some heuristic patterns to group semantic items,
which employ the human-designed syntactic path between two phrases to determine the
relations between any two phrases. In contrast, we collect all the tokens in the depen-
dency path between two phrases as possible patterns. The predicate phraseDepTag and
hasMeanWord are designed to indicate the possible patterns. Note that if these tokens
only contain POS tags dt|in|wdt|to|cc|ex|pos|wp or stop words, the predicate hasMean-
Word is false, otherwise is true. In this way, our system is expected to cover more lan-
guage expressions and answer more questions. Moreover, the SPARQL endpoint is used
7
http://alchemy.cs.washington.edu
8
http://hazy.cs.wisc.edu/hazy/tuffy/
9
http://code.google.com/p/thebeast
1255
Table 2. Descriptions of observed predicates.
Describing the attributes of phrases and relation between two phrases
phraseIndex(p, i, j) The start and end position of phrase p in question.
phrasePosTag(p, pt) The POS tag of head word in phrase p.
phraseDepTag(p, q, dt) The dependency path tags between phrase p and q.
phraseDepOne (p, q) If there is only one tag in the dependency path, the predicate is true.
hasMeanWord (p, q) If there is any one meaning word in the dependency path of two phrases, the predicate is
true.
Describing the attributes of semantic item and the mapping between phrase and semantic item
resourceType(r, rt) The type of semantic item r. Types of semantic items include Entity, Class and Property
priorMatchScore(p, r, s) The prior score of phrase p mapping to semantic item r.
Describing the attributes of relation between two semantic items in knowledge base
hasRelatedness(p, q, s) The semantic coherence of semantic items.
isTypeCompatible(p, q, rr) If semantic items p is type-compatible with semantic items q, the predicate is true.
hasQueryResult(s, p, o, rr1, rr2) If the triple pattern consisting of semantic items s, p, o and relation types rr1, rr2
have query results, the predicate is true.
to verify the type compatibility of two semantic items and whether or not one triple pat-
tern can obtain query results.
The predicate hasRelatedness needs to computes the coherence score between two
semantic items. Following Yahya et al. [9], we use the Jaccard coefficient based on the
inlinks between two semantic items.
The predicate priorMatchScore assigns a score prior to mapping a phrase to a se-
mantic item. We use different ways to compute this scores for different semantic item
types. For entities, we use a normalized score based on the frequencies of a phrase refer-
ring to a entity. For classes and properties, we use different methods. At first, we define
three similarity score metrics as follows: a) s1 : Levenshtein distance score between the
labels of semantic item and phrase; b) s2 : word embedding score, which is the simi-
larity between two phrases, is the maximum value of the cosine of words between two
phrases; c) s3 : instance overlap score, which is computed using the Jaccard coefficient
of instance overlap as a similarity score. The prior scores for mapping phrases to classes
and properties are αs1 + (1 − α)s2 and αs1 + βs2 + (1 − α − β)s3 , respectively. The
parameters are set with empirical values10 .
3.3 Formulas
We use two kinds of formulas for jointly decisions: Boolean and Weighted formulas.
Boolean formulas are hard constraints which must be satisfied with the entire ground
atoms in final inference results. Weighted formulas are soft constraints which could be
violated with some penalties.
1) Boolean Formulas (Hard Constraints) Table 3 lists the boolean formulas used in
this work. The “ ” notation indicates an arbitrary constant. The “||” notation expresses
the number of true grounded atoms in the formula. These formulas express the follow-
ing constraints:
hf1: if a phrase is chosen, then it must have a mapped semantic item;
hf2: if a semantic item is chosen, then its mapped phrase must be chosen;
10
Set α to 0.6 for class, set α and β to 0.3 and 0.3 for property, respectively.
1256
Table 3. Descriptions of boolean formulas.
hf1 hasP hrase(p) ⇒ hasResource(p, )
hf2 hasResource(p, ) ⇒ hasP hrase(p)
hf3 |hasResource(p, )| ≤ 1
hf4 !hasP hrase(p) ⇒!hasResource(p, r)
hf5 hasResource( , r) ⇒ hasRelation(r, , ) ∨ hasRelation( , r, )
hf6 |hasRelation(r1, r2, )| ≤ 1
hf7 hasRelation(r1, r2, ) ⇒ hasResource( , r1) ∧ hasResource( , r2)
hf8 phraseIndex(p1, s1, e1) ∧ phraseIndex(p2, s2, e2) ∧ overlap(s1, e1, s2, e2) ∧
hasP hrase(p1) ⇒!hasP hrase(p2)
hf9 resourceT ype(r, “Entity”) ⇒!hasRelation(r, , “2 1”) ∧ !hasRelation(r, , “2 2”)
hf10 resourceT ype(r, “Entity”) ⇒!hasRelation( , r, “2 1”) ∧ !hasRelation(r, , “2 2”)
hf11 resourceT ype(r, “Class”) ⇒!hasRelation(r, , “2 1”) ∧ !hasRelation(r, , “2 2”)
hf12 resourceT ype(r, “Class”) ⇒!hasRelation( , r, “2 1”) ∧ !hasRelation(r, , “2 2”)
hf13 !isT ypeCompatible(r1, r2, rr) ⇒!hasRelation(r1, r2, rr)
Table 4. Descriptions of weighted formulas.
sf1 priorM atchScore(p, r, s) ⇒ hasP hrase(p)
sf2 priorM atchScore(p, r, s) ⇒ hasResource(p)
sf3 phraseP osT ag(p, pt+) ∧ resourceT ype(r, rt+) ⇒ hasResource(p, r)
sf4 phraseDepT ag(p1, p2, dp+) ∧ hasResource(p1, r1) ∧ hasResource(p2, r2) ⇒
hasRelation(r1, r2, rr+)
sf5 phraseDepT ag(p1, p2, dp+) ∧ hasResource(p1, r1) ∧
hasResource(p2, r2) ∧ !hasM eanW ord(p1, p2) ⇒ hasRelation(r1, r2, rr+)
sf6 phraseDepT ag(p1, p2, dp+) ∧ hasResource(p1, r1) ∧ hasResource(p2, r2) ∧
phraseDepOne(p1, p2) ⇒ hasRelation(r1, r2, rr+)
sf7 hasRelatedness(r1, r2, s) ∧ hasResource( , r1) ∧ hasResource( , r2) ⇒
hasRelation(r1, r2, )
sf8 hasQueryResult(r1, r2, r3, rr1, rr2) ⇒ hasRelation(r1, r2, rr1) ∧
hasRelation(r2, r3, rr2)
hf3: a phrase can map to one semantic item at most;
hf4: if the phrase is not chosen, then its mapping semantic item should not be chosen;
hf5: if a semantic item is chosen, then it should have one relation with other semantic
items at least;
hf6: two semantic items have one relation at most;
hf7: if a relation for two semantic items is chosen, then they must be chosen;
hf8: each two chosen phrases must not overlap;
hf9, hf10, hf11, hf11: the semantic item with type Entity and Class should not have
second argument matching with others;
hf12: The chosen relation for two sematic items must be type-compatible.
2) Weighted Formulas (Soft Constraints) Table 4 lists the weighted formulas used
in this work. The “+” notation in the formulas indicates that each constant of the logic
variable should be weighted separately. Those formulas express the following proper-
ties in joint decisions:
sf1, sf2: The larger the score of phrase mapping to semantic item, the more likely the
corresponding phrase and semantic item should been chosen;
sf3: there are some associations between POS tags of phase and types of mapped se-
mantic item;
sf4, sf5, sf6: there are some associations between the dependency tags in the depen-
1257
dency pattern path of two phases and the types of relation of two mapped semantic
items;
sh7: the larger the relatedness of two semantic items, the more likely they have a rela-
tion;
sf8: if the triple pattern has query results, those semantic items should have correspond-
ing relations.
4 Results and discussion
Stanford dependency parser [15] is used for extracting features from dependency
parse trees. We use the toolkit thebeast11 to learn the weights for the formulas and per-
form MAP inference. The inference algorithm uses a cutting plane approach. And for
parameter learning, we set all initial weights to zeros and use online learning algorithm
with MIRA update rules to update the weights of formulas. The numbers of iterations
for training and testing are set to 10 and 200 epochs, respectively.
Our system could learn the effectiveness patterns. We show the weights of the
learned patterns corresponding with formula sf3 in MLN, as shown in Table 5. From the
table, we can see that nn is mapped to Entity more likely than Class and Property, and
vb is most likely mapped to Property. It proves that our model can learn the effective
and reasonable patterns for QALD.
Table 5. Sample weights of formulas, corresponding with formula sf3.
POS tag of Phrase type of mapped Item Weight
nn Entity 2.11
nn Class 0.243
nn Property 0.335
vb Property 0.517
wp Class 0.143
wr Class 0.025
Table 1 gives the evaluation results with average precision, average recall and F-
measure. It shows the number of question our system can answer, the number of right
and partially right answers among them.
Table 6. Evaluate results on QALD-4 test dataset.
Total Processed Right Partially Recall Precision F-measure
Xser 50 40 34 6 0.71 0.72 0.72
gAnswer 50 25 16 4 0.37 0.37 0.37
CASIA 50 26 15 4 0.40 0.32 0.36
Intui3 50 33 10 4 0.25 0.23 0.24
ISOFT 50 50 10 3 0.26 0.21 0.23
RO FII 50 50 6 0 0.12 0.12 0.12
11
http://code.google.com/p/thebeast
1258
Among the factors that affect performance most are: 1) training set consisting of
110 questions is limited, we found some weights of grounded formula are zero; 2) the
parameters are used for computing prior score mapping phrase to semantic item are
difficult to tune, because we use different method to entities, classes and properties; 3)
lack of global constraints, such as, it is hard to count the number of unmapped tokens
in question.
5 Conclusion
In this paper, we present a question answering system over Linked Data which trans-
lates the natural language questions into the standard RDF data queries (SPARQL). We
present a jointly learning framework for phrase detection, phrases mapping to semantic
items and semantic items grouping. The novelty of our method lies in that we make
joint inference and pattern learning for all subtasks in QALD by using first-order logic.
Our benchmark results demonstrate the effectiveness of the proposed method.
References
1. C. Bizer, T. Heath, and T. Berners-Lee, “Linked data-the story so far,” International journal
on semantic web and information systems, vol. 5, no. 3, pp. 1–22, 2009.
2. S. Auer, C. Bizer, G. Kobilarov, J. Lehmann, R. Cyganiak, and Z. Ives, “Dbpedia: A nucleus
for a web of open data,” in The semantic web. Springer, 2007, pp. 722–735.
3. K. Bollacker, C. Evans, P. Paritosh, T. Sturge, and J. Taylor, “Freebase: a collaboratively
created graph database for structuring human knowledge,” in SIGMOD, 2008.
4. F. M. Suchanek, G. Kasneci, and G. Weikum, “Yago: a core of semantic knowledge,” in
WWW, 2007.
5. S. Walter, C. Unger, P. Cimiano, and D. Bär, “Evaluation of a layered approach to question
answering over linked data,” in The Semantic Web–ISWC 2012. Springer, 2012, pp. 362–
374.
6. V. Lopez, V. Uren, M. Sabou, and E. Motta, “Is question answering fit for the semantic web?:
a survey,” Semantic Web, vol. 2, no. 2, pp. 125–155, 2011.
7. V. Lopez, E. Motta, and V. Uren, “Poweraqua: Fishing the semantic web,” in The Semantic
Web: research and applications. Springer, 2006, pp. 393–410.
8. C. Unger, L. Bühmann, J. Lehmann, A.-C. Ngonga Ngomo, D. Gerber, and P. Cimiano,
“Template-based question answering over rdf data,” in WWW, 2012.
9. M. Yahya, K. Berberich, S. Elbassuoni, M. Ramanath, V. Tresp, and G. Weikum, “Natural
language questions for the web of data,” in EMNLP, 2012.
10. M. Richardson and P. Domingos, “Markov logic networks,” Machine learning, vol. 62, no.
1-2, pp. 107–136, 2006.
11. N. Nakashole, G. Weikum, and F. Suchanek, “Patty: a taxonomy of relational patterns with
semantic types,” in EMNLP, 2012.
12. A. Fader, S. Soderland, and O. Etzioni, “Identifying relations for open information extrac-
tion,” in EMNLP, 2011.
13. J. Berant, A. Chou, R. Frostig, and P. Liang, “Semantic parsing on freebase from question-
answer pairs,” in EMNLP, 2013.
14. S. Riedel, “Improving the accuracy and efficiency of map inference for markov logic.” UAI,
2008.
15. M.-C. De Marneffe, B. MacCartney, C. D. Manning et al., “Generating typed dependency
parses from phrase structure parses,” in LREC, 2006.
1259