<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>A Hybrid Similarity Computing Method for KBQA</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Chunpei Wang</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Xiaowang Zhang</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Zhiyong Feng</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>College of Intelligence and Computing, Tianjin University</institution>
          ,
          <addr-line>Tianjin 300350</addr-line>
          ,
          <country country="CN">China</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>With the rapid growth of knowledge bases(KBs), knowledgebase question answering has drawn huge attention in recent years. Most existing KBQA methods translate questions into SPARQLs to help endusers access the knowledge base represented by RDF more naturally. However, a natural language question is always corresponding to multiple candidate SPARQLs due to the gap between the unstructured Question and the structural SPARQL query. To pick the best SPARQL query from the candidate SPARQL query set, in this Poster, we propose a hybrid similarity computing method to rank the SPARQL query. Firstly, we employ two attentive recurrent neural networks to capture the semantic Similarity between the SPARQL query and the Question. Secondly, we compute the string similarity between the SPARQL query and the Question by leveraging the convolutional neural network. Our method can capture the two-level Similarity between the Question and the SPARQL. Experiments show that our method can improve the effectiveness of KBQA.</p>
      </abstract>
      <kwd-group>
        <kwd>KBQA Semantic Parsing RDF SPARQL</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>Generally, we will obtain multiple SPARQL queries after the semantic parsing
stage since the ambiguity between the natural language question and the
knowledge base. For instance, the entity mention St. Lawrence of the Question \What
body of water does St.Lawrence ow into? " will be mapped to a set of semantic
instances in the KB, e.g., E = fhSiant Lawrencei, or hSiant Lawrence Riverig.
Thus, the main challenge in the semantic parsing stage is how to pick the best
SPARQL query in the candidate query set.</p>
      <p>Most existing KBQA work maps the Question and the KB facts(triple) to
a common embedding space. The Similarity between the question vector and
the SPARQL vectors can be conveniently computed. However, these methods
tend to lose original word interaction information. To preserve more original
information, we propose a hybrid similarity computing method to pick the best
SPARQL query from the candidate set. Both consider the semantic Similarity
and the strong Similarity between the Question and the SPARQL query.</p>
      <p>Copyright 2020 for this paper by its authors. Use permitted under Creative
Commons License Attribution 4.0 International (CC BY 4.0).</p>
    </sec>
    <sec id="sec-2">
      <title>Our Approach</title>
      <p>In this Poster, we present a hybrid ranking model to rank the SPARQL query,
which considers both string similarity and Semantic Similarity. Given the
natural language question N , for each query qi in the candidate SPARQL set, we
compute the similarity score S(N; qi) that represents the semantic Similarity
between N and qi. Finally, all candidate queries are ranked via their similarity
scores with N .</p>
      <p>Semantic-level Similarity. We construct an attentive recurrent neural
network for computing semantic-level similarity between question N and the
candidate query qi. The model uses an encoder-compare framework which encodes
the semantic information of N and qi into high-dimensional embedding space
and then estimates their similarity via multilayer perceptron(MLP).
{ Encoding. Firstly, each elements in question N and query qi is mapped to
its corresponding embedding vectors fw1; : : : ; wLg, where L is the length
of the question or query. And then all the embeddings will be input into
a bidirectional GRUs neural network to learn the hidden representations
H1:L = [h1; : : : ; hL], where hi is the concatenation of forward and backward
vectors learned at time i. Since each word contribute di erently to the full
sentence semantics, the model would pay di erent attention to each word
and learn promising vectors to represent the question/query sentences. The
self-attention model is used here to learn the weight of each word semantic
for the input sentence. The semantic representation Y of sentence can be
calculated as follows:</p>
      <p>Y =</p>
      <p>L
X aihi;
i=1
a = Attention(Q; K; V ) = sof tmax</p>
      <p>QK&gt;
pdk</p>
      <p>
        V
(
        <xref ref-type="bibr" rid="ref1">1</xref>
        )
(
        <xref ref-type="bibr" rid="ref2">2</xref>
        )
where fQ; K; V g are the shorthand for fquery, key, valueg, which are three
matrices that mapped with the same input. K and V is a one-to-one
correspondence with key-value relation. Q could be the hidden state to be
processed, such as hi. First, the dot product between Q and K is computed,
which will be divided by a scale factor of pdk to prevent the result from
being too large. And then, the result will be processed with softmax
function to get normalized probability, which will be multiplied by V to get the
weight. Finally, each hidden presentation hi is multiplied by the attention
weights and summed to get semantic representation Y .
{ Similarity estimation. With the representations Yp and Yqi of question p and
query qi, their similarity will be calculated by a MLP layer
z1 = f W &gt; [Yp; Yqi ] + b
(
        <xref ref-type="bibr" rid="ref3">3</xref>
        )
where W is the parameters to be learned, b is the deviation and f ( ) is
an activation function. The semantic information extracted from the two
sentences are spliced as the input of MLP hidden layer, which nonlinearly
mapped the two sentences into their Similarity.
      </p>
      <p>String-level Similarity. The Similarity over string-level is evaluated via a
text-matching model. Some words or phrases with the same meaning may be
expressed di erently in the Question and query, i.e., the pair of words (musical; music)
have similar semantics. Since the high-level semantic embedding cannot preserve
these words interaction information, we construct a similarity matrix whose
elements represent the similarities between question words and query words,
regarding it as a two-dimensional vector space to utilize the convolutional layer to
capture the matching features.</p>
      <p>
        { Similarity matrix. Firstly, we construct a similarity matrix M , where each
element Mij indicates the basic interaction. The Mij can be calculated as:
Mij = ui
vi
Where ui and vj denotes the i-th and the j-th word was embedding in
Question and query, respectively. The operator stands for a general operator to
compute the Similarity. Here the matrix can obtain the Similarity of words
with di erent expressions.
{ Convolution layer. The di erent levels of matching patterns can be extracted
by a convolutional kernel. The k-th kernel wk scans over the similarity matrix
M to generate a feature map gk:
(
        <xref ref-type="bibr" rid="ref4">4</xref>
        )
(
        <xref ref-type="bibr" rid="ref5">5</xref>
        )
(6)
(7)
k
gi;j =
rk 1 rk 1
X X wsk;t
      </p>
      <p>Combination. With three feature (z1; z2; z3) generated from two level
similarity, we utilize a linear layer to learn their respective contribution for holistic
similarity score:</p>
      <p>S(p; qi) = Sigmoid W &gt;[z1; z2; z3] + b :
(8)
Finally, all candidate queries qi will be sorted with the similarity scores S(p; qi).</p>
    </sec>
    <sec id="sec-3">
      <title>Experiments and Evaluation</title>
      <p>Our Method
BiCNN(Yih et al.)
AMPCNN (Wenpeng et al.)
HR-BiLSTM (Yu et al.)
Multiple View Matching (Yu et al.)</p>
      <p>SimpleQuestions is a single-relation KBQA dataset. This dataset consists
of questions annotated with a corresponding fact from Freebase that provides
the answer. We report Accuracy as previous studies. We verify our proposed
approach on the SimpleQuestion dataset. Table 1 summarizes the experimental
results of di erent methods on answer selection and knowledge base question
answering. Clearly, our method achieves the state-of-the-art results in
SimpleQuestions, which con rms the e ectiveness of our solution.</p>
    </sec>
    <sec id="sec-4">
      <title>Acknowledgments</title>
      <p>This work is supported by the National Key Research and Development Program
of China (2017YFC0908401) and the National Natural Science Foundation of
China (61972455). Xiaowang Zhang is supported by the Peiyang Young Scholars
in Tianjin University (2019XRX-0032).</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>A. P. B. Veyseh</surname>
          </string-name>
          , \
          <article-title>Cross-lingual question answering using common semantic space,"</article-title>
          <source>in Proceedings of TextGraphs@NAACL-HLT</source>
          <year>2016</year>
          , pp.
          <volume>15</volume>
          {
          <issue>19</issue>
          ,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <given-names>L.</given-names>
            <surname>Dong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Wei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Zhou</surname>
          </string-name>
          , and
          <string-name>
            <given-names>K.</given-names>
            <surname>Xu</surname>
          </string-name>
          , \
          <article-title>Question answering over Freebase with multi-column convolutional neural networks</article-title>
          in
          <source>ACL 2015, July 26-31</source>
          ,
          <year>2015</year>
          , Beijing, China, Volume
          <volume>1</volume>
          :
          <string-name>
            <given-names>Long</given-names>
            <surname>Papers</surname>
          </string-name>
          , pp.
          <volume>260</volume>
          {
          <issue>269</issue>
          ,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <given-names>A.</given-names>
            <surname>Bordes</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Chopra</surname>
          </string-name>
          , and
          <string-name>
            <given-names>J.</given-names>
            <surname>Weston</surname>
          </string-name>
          , \
          <article-title>Question answering with subgraph embeddings,"</article-title>
          <source>in EMNLP 2014, October 25-29</source>
          ,
          <year>2014</year>
          , Doha,
          <string-name>
            <surname>Qatar,</surname>
          </string-name>
          <article-title>A meeting of SIGDAT, a Special Interest Group of the ACL</article-title>
          , pp.
          <volume>615</volume>
          {
          <issue>620</issue>
          ,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <given-names>X.</given-names>
            <surname>Yao</surname>
          </string-name>
          and
          <string-name>
            <given-names>B. V.</given-names>
            <surname>Durme</surname>
          </string-name>
          , \
          <article-title>Information extraction over structured data: Question answering with freebase,"</article-title>
          <source>in ACL 2014</source>
          , June 22-27,
          <year>2014</year>
          , Baltimore,
          <string-name>
            <surname>MD</surname>
          </string-name>
          , USA, Volume
          <volume>1</volume>
          :
          <string-name>
            <given-names>Long</given-names>
            <surname>Papers</surname>
          </string-name>
          , pp.
          <volume>956</volume>
          {
          <issue>966</issue>
          ,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <given-names>X.</given-names>
            <surname>Lu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Pramanik</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. S.</given-names>
            <surname>Roy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Abujabal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wang</surname>
          </string-name>
          , and G. Weikum, \
          <article-title>Answering complex questions by joining multi-document evidence with quasi knowledge graphs,"</article-title>
          <source>in Proceedings of the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval</source>
          ,
          <string-name>
            <surname>SIGIR</surname>
          </string-name>
          <year>2019</year>
          , Paris, France,
          <source>July 21-25</source>
          ,
          <year>2019</year>
          ., pp.
          <volume>105</volume>
          {
          <issue>114</issue>
          ,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>