<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Improving Knowledge Base Question Answering with Question Understanding Augment</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Peiyun</forename><surname>Wu</surname></persName>
							<email>wupeiyun@tju.edu.cn</email>
							<affiliation key="aff0">
								<orgName type="department">College of Intelligence and Computing</orgName>
								<orgName type="institution">Tianjin University</orgName>
								<address>
									<postCode>300350</postCode>
									<settlement>Tianjin</settlement>
									<country key="CN">China</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Xiaowang</forename><surname>Zhang</surname></persName>
							<email>xiaowangzhang@tju.edu.cn</email>
							<affiliation key="aff0">
								<orgName type="department">College of Intelligence and Computing</orgName>
								<orgName type="institution">Tianjin University</orgName>
								<address>
									<postCode>300350</postCode>
									<settlement>Tianjin</settlement>
									<country key="CN">China</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Improving Knowledge Base Question Answering with Question Understanding Augment</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">2AB22E6AC2CFEA6FDA652A0998185369</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T08:34+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>The basement of knowledge base question answering (KBQA) is to understand the given question and extract the meaning from it. Existing works largely focus on generating query graphs to represent the semantics of the question while ignoring understand the real meaning of it. To augment question understanding, in this paper, we leverage rich external linguistic knowledge to enhance question semantics. First, we integrate the sememe and gloss information into words, where sememe (the minimum semantic units of word meanings) and gloss (sense definition) are used to disambiguate the word sense and enrich questions information. Moreover, we present a co-attention network to build co-dependent representations for the sememe and gloss. Experiments evaluated on two data sets show that our model outperforms existing approaches.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>Semantic parsing is an important approach to KBQA, which constructs a query structure (called query graph) that represents the semantics of questions. Semantic parsing based approaches effectively transform questions into logical forms where the reliability of logical forms can ensure the correctness of answering questions. The success of semantics parsing lies in representing the semantics of questions to better capture users' intention.</p><p>However, in recent years, many semantic parsing approaches focus on complex query graph generation and re-ranking <ref type="bibr" target="#b0">[1,</ref><ref type="bibr" target="#b1">2,</ref><ref type="bibr" target="#b4">5]</ref> with paying little attention to understanding the meanings of questions accurately. They aim to leverage a ranking model to score and find the best query graph. As a result, existing works without processing ambiguous questions cannot always perform query graph ranking better.</p><p>In this paper, we propose an augmented question representation method by leveraging the sememe and gloss information. Specifically, we take advantage of the gloss information from WordNet and the sememe <ref type="bibr" target="#b2">[3]</ref> information from Hownet into word embeddings from the given question. A word may have multiple senses and a sense consists of several sememes and a gloss. ...  </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Sense selector Gloss selector</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Co-Attention</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Our Approach</head><p>Given a question Q = {w 1 , . . . , w n }, we generate its candidate query graph set by method in <ref type="bibr" target="#b0">[1]</ref>. We measure the semantic similarities between the question and each query graph to find the optimal one. For word w i in Q, we denote the set of its senses as S wi , and E wi = {e i1 , . . . , e ik } represents an unordered set of all sememes contained in w i . We assume for word w i that we have a gloss set G wi . Our model is shown in Fig. <ref type="figure" target="#fig_1">1</ref>.</p><p>Sense Selector: This selector selects each word sense that is most relevant to the context. We first feed Q into a bi-directional long short-term memory network (Bi-LSTM) to generate the hidden representation {h 1 , . . . , h n }. Then we generate the context representation of Q as below:</p><formula xml:id="formula_0">context = n i=1 softmax(tanh(h i • ( 1 n n i=1 h i ))) • h i<label>(1)</label></formula><p>To calculate the correlation between each sememe and context, we use Sigmoid function to obtain probability value by:</p><formula xml:id="formula_1">p(e ij |context) = Sigmoid context • e ij , ∀ j ∈ (1, . . . , k)<label>(2)</label></formula><p>For each sense in S wi , its probability is calculated by the average of all the sememe probabilities that it contains. In this way, we can select the sense with the highest probability value under the current context and denote it as S wi max = {e 1 , . . . , e k }, where {e 1 , . . . , e k } represents all sememe vectors it contains.</p><p>Gloss Selector: This selector selects the gloss that is most relevant to the selected word sense. Analogously, we use Sigmoid function to obtain the highest probability of gloss that is most relevant to the average embedding of S wi max . We denote the selected gloss as G wi max = {o 1 , . . . , o m }, where {o 1 , . . . , o m } represents all word embeddings it contains.</p><p>Co-Attention: To model the mutual influence and highlight the important information in the sememe and gloss, we introduce a co-attention network to dynamically combine the sememe and gloss representation as:</p><formula xml:id="formula_2">U s = tanh(S wi max • G wi max ), U g = tanh(G wi max • S wi max )<label>(3)</label></formula><formula xml:id="formula_3">SG wi = λ k j=1 [U s • softmax(U s : )] :j + (1 − λ) m j=1 [U g • softmax(U g : )] :j<label>(4)</label></formula><p>where SG wi is the combination representation of the selected sememe and gloss of word w i , λ is the parameter and λ ∈ [0, 1]. softmax(U s : ) and softmax(U g : ) are attention weight matrix for softmax function across each column of U s and U g , respectively. [] :j denotes j-th column of []. Finally, we concatenate SG wi to the word embedding of w i to enrich the semantics and reduce ambiguity.</p><p>Question Representation: We treat {x 1 , . . . , x n } as the initial representations of the given question which has been integrated with sememe and gloss. To further augment the contextual embeddings of the question, we parse the question into its syntactic dependencies graph D G and adopt relational graph convolutional network (RGCN) to digest this structural information:</p><formula xml:id="formula_4">x (l+1) i = ReLU   r∈R j∈N r i 1 |N r i | W<label>(l) r x (l) j + W (l) 0 x (l) i </label></formula><formula xml:id="formula_5"><label>(5)</label></formula><p>Here R is a set of dependency-relation. l is l-th layer, N r i is the set of all rneighbors of i-th node in D G . Note that W 0 and W r are weighted matrixes. Finally, we apply a pooling operation after the last RGCN layer to get the representation of the question.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Relation Representation</head><p>We represent relations from different granularity in a query graph. For each relation, we take its relation-level and word-level representations into consideration. The word-level relation is calculated by its average word embeddings. The relation-level representation is the vector of unique token of relation name. Then, each relation is represented by the sum operation and we perform max pooling over relations to obtain the final relation representation.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Experiments and Evaluations</head><p>We use Wikidata as our KB and conduct on two data sets, namely, WebQSP-WD(WSPWD) <ref type="bibr" target="#b0">[1]</ref> and QALD-7 (Task 4, English), both support for Wikidata. We use F1-score as metrics, where all results are macro-averaged scores.</p><p>By Table <ref type="table" target="#tab_1">1</ref>, we show that our model outperforms all datasets. Our model achieves 54.2%, 23.3%, 8.9%, 11.9% higher F1-score compared to STAGG, HR-BiLSTM, GGNN, Slot-Matching on WSPWD. Analogously, we achieve 59.3%, 45.7%, 39.1%,21.7% higher F1-score on QALD-7. We observe that if we only integrate sememe information "+sememe" or gloss information "+gloss", our model performs worse but still keep competitive. We can conclude the effectiveness of our augmented question representation with sememe and gloss integration.  To measure the performance across questions of different complexity, we break down the performance by the number of relations that are needed to find the correct answer on WSPWD, and results are shown in Fig. <ref type="figure" target="#fig_2">2</ref>, we can see that our model is effective in dealing with different complexity questions.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Conclusion</head><p>In this paper, we augment question understanding to improve KBQA. The sememe and gloss information can benefit from each other and enhance question semantics. In this way, our approach provides a new method of usage of external knowledge in question representation. In future work, we are interested in extending our model for more complex practical questions.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Fig. 1 :</head><label>1</label><figDesc>Fig. 1: Diagram for Our Model.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Fig. 2 :</head><label>2</label><figDesc>Fig. 2: The number of relations to find correct answers</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head></head><label></label><figDesc>To highlight important information in the sememe and gloss, we present a co-attention network to generate better representations.</figDesc><table><row><cell>...</cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell></cell><cell></cell><cell>context</cell><cell></cell><cell></cell></row><row><cell>1.the person who holds the</cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell>office of head of state of the</cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell>2.an executive officer of a firm United States government</cell><cell>president (state president)</cell><cell>president (principal)</cell><cell>...</cell><cell>president (CEO)</cell></row><row><cell>or corporation</cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell>3.the head administrative officer of a college or</cell><cell>human/occupation/ politics/HeadOfStat</cell><cell>human/occupation /education/official/</cell><cell>...</cell><cell>human/occupation /economy/official/</cell></row><row><cell>university</cell><cell>e/manage/country...</cell><cell>study/teach...</cell><cell></cell><cell>manage/primary...</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 1 :</head><label>1</label><figDesc>Overall Average Results over</figDesc><table><row><cell>Wikidata</cell><cell></cell></row><row><cell>Model</cell><cell>WSPWD QALD-7</cell></row><row><cell>STAGG(2015) [2]</cell><cell>0.1828 0.1861</cell></row><row><cell cols="2">HR-BiLSTM(2017) [5] 0.2287 0.2035</cell></row><row><cell>GGNN(2018)[1]</cell><cell>0.2588 0.2131</cell></row><row><cell cols="2">Slot-Matching(2019)[4] 0.2519 0.2436</cell></row><row><cell>+sememe</cell><cell>0.2459 0.2546</cell></row><row><cell>+gloss</cell><cell>0.2597 0.2743</cell></row><row><cell>Our</cell><cell>0.2819 0.2965</cell></row></table></figure>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Acknowledgments</head><p>This work is supported by the National Key Research and Development Program of China (2017YFC0908401) and the National Natural Science Foundation of China (61972455). Xiaowang Zhang is supported by the Peiyang Young Scholars in Tianjin University (2019XRX-0032).</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Modeling semantics with gated graph neural networks for knowledge base question answering</title>
		<author>
			<persName><forename type="first">D</forename><surname>Sorokin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Gurevych</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">COLING&apos;</title>
				<imprint>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page" from="3306" to="3317" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Semantic parsing via staged query graph generation: question answering with knowledge base</title>
		<author>
			<persName><forename type="first">W</forename><surname>Yih</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Chang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>He</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Gao</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">ACL&apos;2015</title>
				<imprint>
			<biblScope unit="page" from="1321" to="1331" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">A set of postulates for the science of language</title>
		<author>
			<persName><forename type="first">L</forename><surname>Bloomfield</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Language</title>
		<imprint>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="issue">3</biblScope>
			<biblScope unit="page" from="153" to="164" />
			<date type="published" when="1926">1926</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Learning to rank query graphs for complex question answering over knowledge graphs</title>
		<author>
			<persName><forename type="first">G</forename><surname>Maheshwari</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Trivedi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Lukovnikov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Chakraborty</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Fischer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Lehmann</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">ISWC&apos;19</title>
				<imprint>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="487" to="504" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Improved neural relation detection for knowledge base question answering</title>
		<author>
			<persName><forename type="first">M</forename><surname>Yu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Yin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><forename type="middle">S</forename><surname>Hasan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">N</forename><surname>Santos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Xiang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Zhou</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">ACL&apos;2017</title>
				<imprint>
			<biblScope unit="page" from="571" to="581" />
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
