<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">A cross-domain natural language interface to databases using adversarial text method</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author role="corresp">
							<persName><forename type="first">Wenlu</forename><surname>Wang</surname></persName>
							<email>wenluwang@auburn.edu</email>
							<affiliation key="aff0">
								<orgName type="institution">Auburn University</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Wei-Shinn</forename><surname>Ku</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">Auburn University</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Haixun</forename><surname>Wang</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">Auburn University</orgName>
							</affiliation>
						</author>
						<title level="a" type="main">A cross-domain natural language interface to databases using adversarial text method</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">D24FD0DDA93B4D5E0FE40F228713526C</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-25T08:34+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>A natural language interface (NLI) to databases is an interface that supports natural language queries to be executed by database management systems (DBMS). However, most NLIs are domain specific due to the complexity of the natural language questions, and an NLI trained on one domain is hard to be transferred another due to the discrepancies between di↵erent ontology. Inspired by the idea of stripping domain-specific information out of natural language questions, we propose a cross-domain NLI with a general purpose question tagging strategy and a multi-language neural translation model. Our question tagging strategy is able to extract the "skeleton" of the question that represents its semantic structure for any domain. With question tagging, every domain will be handled equally with a single multi-language neural translation model. Our preliminary experiments show that our multi-domain model has excellent cross-domain transferability.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">INTRODUCTION</head><p>Relational databases are widely adopted in real-world applications <ref type="bibr" target="#b14">[15,</ref><ref type="bibr" target="#b13">14]</ref>. However, it requires a certain knowledge of query languages to operate on DBMSs, which motivated the study of NLI to databases <ref type="bibr" target="#b0">[1]</ref> with the purpose of making DBMSs operable by anyone without training.</p><p>The challenges of NLI to databases lies in the discrepancies of di↵erent ontology, which makes general-purpose NLI hard to achieve. Most existing general purpose NLIs exploit syntax-guided decoding and require the grammar of the structured queries (domain specific grammar) as part of the model. Such a model cannot be shared between di↵erent grammars, while we propose a general purpose model where di↵erent types of queries and di↵erence domains share the same components. To overcome the obstacles of generalizing one NLI model to di↵erent domains or even unseen domains, we perform a pre-processing step inspired by the strategy of separating domain-specific information from the question <ref type="bibr" target="#b21">[22]</ref>. By detaching domain-specific data elements, the NLI model is able to focus on the semantic meaning and agnostic of the natural language question, which facilitates the cross-domain generalization.</p><p>We first "strip" a natural language question (shown in Figure <ref type="figure">1</ref>), each type of the query (SQL and Lambda expression in our examples) is treated equally, then translate the tagged question to a structured query. The definition of "strip" is enclosing a phrase that describes a data element (tables, columns, values, keywords, etc.) appearing in the query by inserting a symbol (k1, v1, etc.) representing the type and index of the data element in front of the phrase, and an "end of element" symbol (e.g., heoei) at the end of the phrase. In Figure <ref type="figure">1</ref>, we show two types of queries (Lambda Expression and SQL), k represents a column field, a table name, or a keyword, and v represents a value. </p><p>Figure <ref type="figure">1</ref>: Two types of queries (SQL and Lambda expression) with corresponding Natural language questions.</p><p>Another challenge of our cross-domain task is to handle di↵erent query types. The aforementioned symbol insertion strategy is able to handle questions of di↵erent types equally but fails to di↵erentiate them. Inspired by Google's multilingual translation model <ref type="bibr" target="#b10">[11]</ref> where an artificial token is introduced at the beginning of the input sentence to indicate the target language. We prefix a query type symbol to indicate the target query type the NLI model should covert to (e.g., hSQLi, hlambdai). For instance, consider the following question -&gt; SQL pair:</p><p>Which is the highest score? -&gt; SELECT MAX(score) It will be modified to:</p><p>hSQLi Which is the highest score? -&gt; SELECT MAX(score) Such an approach only needs to prefix one additional token, we will validate in the preliminary experiments that it is the simplest but e↵ective approach.</p><p>The core design of our symbol insertion strategy lies in how to identify the phrase that describes a data element appears in the corresponding query. The phrase might not the data element "release date" is described as "to release on" in the question. Inspired by gradient-based adversarial text method, we propose an adversarial method towards a data element detector. Given a natural language question q and a data element e, the data element detector will predict whether e is mentioned in q.</p><p>Figure <ref type="figure">1</ref> presents two examples. In example (1), the question (imply natural language question in this paper) is converted to a lambda expression, and example (2) is converted to a SQL query.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">ADVERSARIAL TEXT METHOD</head><p>It has been demonstrated that adding a carefully crafted small noise is able to fool the deep neural network models into wrong predictions, while the small noise makes unnoticeable visual di↵erence <ref type="bibr" target="#b3">[4]</ref>. Most of the adversarial attack methods on the text <ref type="bibr" target="#b11">[12,</ref><ref type="bibr" target="#b18">19,</ref><ref type="bibr" target="#b9">10]</ref> try to perturb the features (e.g., words, characters, and phrases) that are the most influential on the predictions. Inspired by gradient-based adversarial text attacks <ref type="bibr" target="#b4">[5]</ref>, we propose our own solution to identify the position of a data element in a question.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">DESIGN</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">Overview</head><p>Given a (question, query) pair, our core methodology is to insert pre-designed symbols and enclose data elements mentioned in the question to achieve the purpose of handling every sample (of di↵erent domain/type) equally. Figure <ref type="figure" target="#fig_0">2</ref> shows the framework of our approach bottom-up.</p><p>1. Build a binary classifier BC as a data element detector to predict whether a data element e appears in a question q's corresponding query p by the semantic meaning of the question, which takes q and e as inputs without referencing p.</p><p>2. Inspired by <ref type="bibr" target="#b4">[5]</ref>, we search for the most influential phrase in the question using gradient-based adversarial text methods. We refer "the most influential phrase" as the phrase that describes the data element e theoretically.</p><p>3. We insert symbols in q to enclose the phrases that describes the data elements, denoted as q 0 . Since a query type symbol (e.g., hSQLi) is prefixed to q 0 , a user is able to select a desired query type.</p><p>4. Build a multi-lingual cross-domain sequence-to-sequence (seq2seq) model to translate q 0 to p 0 , and p 0 is a query where the data elements are replaced by symbols inserted in q.</p><p>5. Inserted symbols are replaced with data elements to form the original query (convert p 0 back to p).</p><p>Figure <ref type="figure" target="#fig_0">2</ref> shows two examples of (q, q 0 , p 0 , p) correspond to Figure <ref type="figure">1</ref>. In Figure <ref type="figure">1</ref>(1), data elements "city" and "Virginia" are able to be detected by comparing against the database using string match directly. So are "movie" and "May 19 2019" in Figure <ref type="figure">1</ref>(2). However, detecting data elements "location" and "release date" is problematic, so we use the pre-trained binary classifier BC. If BC is well trained, it will produce positive predictions for data element e = ["location","release date"]. As the true label, p is not involved in the process. After translating q 0 to p 0 , where data elements are represented as symbols, we perform the final step of converting p 0 back to p. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>&lt;SQL&gt; Which &lt;k1&gt; movies &lt;eoe&gt; were scheduled &lt;k2&gt; to release on &lt;eoe&gt; &lt;v2&gt;</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2">Data Element Search</head><p>Data elements include table names, column fields, and column values in databases, and keywords in query grammar. For example, "movies", "release date", and "May 19 2019" are data elements in SQL query "SELECT movie WHERE release date EQUAL May 19 2019", and the other elements SELECT WHERE EQUAL belong to the template of SQL sketch. In "city(A),location(A,B),const(B, stateid(Virginia))", "city" is a table name, "stateid" is a column field, "Virginia" is a column value, and "location" is a keyword in lambda expression grammar. The challenge is to discover all the data elements from the question , and for symbol insertion purpose, we need to discover which phrase describes each data element. With such symbols insertion, we relieve the burden of identify the data elements from seq2seq model, and makes it focus on learning semantic structure of the question and logic of the data elements.</p><p>We have two challenges to tackle for data element search (use Figure <ref type="figure" target="#fig_0">2</ref> as an example):</p><p>1. Identify whether a data element is described in the question. Ultimately, we are trying to detect all the data elements that constitute the query, and we have to infer those data elements from the natural language question based on its semantics. In q1, we need to identify all the data elements that are described in the question (e.g., "city", "location", and "Virginia"). In q2, we need to identify "release data", "movie", and "May 19 2019".</p><p>2. The phrase that describes a data element needs to be identified by its semantic meaning and contextual comprehension. In q1 of Figure <ref type="figure" target="#fig_0">2</ref>, "located in" is identified as the most influential phrase while describing the key word "location". In q2, "to release on" is identified as the phrase that describes "release date".</p><p>To address these two challenges, we propose our general purpose data element search strategy with two steps:</p><p>• We propose a Data Element Detector (Sec 3.2.1) for the first challenge, which is a binary classifier with a question q and data element e as an input. The detector is able to detect whether the data element e is mentioned in question q. As presented in Figure <ref type="figure" target="#fig_0">2</ref>, a Data Element Detector is shared among all the domains.</p><p>• In the case of positive prediction in the previous step, we propose an Adversarial Text Method (Sec 3.2.2) for the second challenge, which relies on the information learned by the binary classifier from the first step.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.1">Data Element Detector</head><p>We use a bi-directional attentive recurrent neural network to achieve question understanding. For a question q composed of n tokens [q1, ..., qn] and a data element e composed of m tokens [e1, ..., em], we use a pre-trained Glove embedding to initiate a word embedding layer. On top of the embedding layer, we use LSTM cells to produce hidden states for each time step (each word in q). We build a similar structure for e. We denote the top layer hidden states as</p><formula xml:id="formula_1">h q = [h q 1 , • • • , h q n ] h e = [h e 1 , • • • , h e m ]</formula><p>We build a bi-directional LSTM layer on h q with attention over h e .</p><formula xml:id="formula_2">! d0 = 0 ! t =  h e t ! t ! dt = LSTM! ⇣ ! t , ! dt 1 ⌘ ! etj = v T Tanh(W0h q j + W1h e + W2 ! dt 1 ) ! ↵tj = ! etj / X j 0 ! e tj 0 ! t = n X j=1 ! ↵tj h q j</formula><p>where W0, W1, W2 ,and v are model parameters. Here t enumerates each time step for e, and j enumerates each token in q. We compute bi-directional output dt = [ ! dt , dt ], and feed it to a multi-layer perception for binary prediction.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.2">Adversarial Text Method</head><p>With the adversarial text method, given a data element e that has a positive prediction from the binary classifier, we propose to search a phrase of the question that describes e. We describe our adversarial text method as follows.</p><p>1. We have trained a Data Element Detector that takes a question q and a data element e as inputs and predicts whether e is described in q.</p><p>2. We search for the most influential phrase in q using gradient-based adversarial text methods <ref type="bibr" target="#b4">[5]</ref>. There are three possible directions (the loss gradient of the Data Element Detector with q and e as inputs is rL(q, e)):</p><p>-DeepFool <ref type="bibr" target="#b12">[13]</ref>. We iteratively search the optimal direction in which only a minimum perturbation is needed to a↵ect the prediction. Theoretically, the optimal direction is f (q,e) ||rL(q,e)|| 2 2 rL(q, e) where f (•) denotes the Data Element Detector.</p><p>-Fast Gradient Method FGM <ref type="bibr" target="#b5">[6]</ref>. We add a noise that is proportional to either rL(q, e) or sign(rL(q, e)) to the original sample to change the prediction of the Data Element Detector. In particular, the noise for each token qi is proportional to @L(q,e) @q i .</p><p>-JSMA <ref type="bibr" target="#b15">[16]</ref>. We calculate the Jacobian-based saliency map based on rL(q, e), and perturb one token at a time. The chosen token has the highest saliency value. Since all the methods are trying to add minimum noise that influences the prediction the most. The locations where the noise is added will be the positions of tokens that constitute the most influential phrase -i,e., the phrase that describes the data element e.</p><p>3. We search for a phrase in the question where adding a small perturbation will a↵ect the prediction dramatically.</p><p>The challenge of our adversarial text method is the discreteness of the text domain. Words or characters are discrete variables thus indi↵erentiable. To overcome such problem, we propose to calculate the loss gradient (rL) of the target model w.r.t. the embedding of each word, where the embedding space is smooth.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3">Neural Machine Translation</head><p>We denote an question post symbol insertion as q 0 and the corresponding query post symbol replacement as p 0 . We train a seq2seq model to translate q 0 to p 0 :</p><formula xml:id="formula_3">p 0 = seq2seq(q 0 )</formula><p>Encoder is a stacked bi-directional multi-layer recurrent neural network (RNN). Decoder is a one-layer attentive RNN.</p><p>We use a single multilingual neural translation model for our cross-domain NLI task. We believe with a prefixed query type symbol (e.g., hSQLi), a multi-domain model is able to handle di↵erent query types, and each query type is treated equally.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">RELATED WORK</head><p>NLI to databases was first formally introduced in <ref type="bibr" target="#b0">[1]</ref>. Semantic parsing <ref type="bibr" target="#b16">[17,</ref><ref type="bibr" target="#b22">23,</ref><ref type="bibr" target="#b8">9]</ref> and cross-domain semantic parsing <ref type="bibr" target="#b6">[7,</ref><ref type="bibr" target="#b19">20]</ref> are applied in NLI to databases. However, due to the incompatibility among di↵erent domains, cross-domain task remains unsolved. The sketch-based solutions are also extensively studied, which is first proposed in <ref type="bibr" target="#b23">[24]</ref>. A deep model is trained to fill the slots in the sketch. An extension of sketch-based solution <ref type="bibr" target="#b25">[26]</ref> relies on a knowledge base to identify the column values. Such a strategy is confined in pre-defined sketch and existing knowledge base. seq2seq model are also exploited to serve as a translator <ref type="bibr" target="#b7">[8,</ref><ref type="bibr" target="#b27">28]</ref>, which has no limitations on query sketch.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">PRELIMINARY EXPERIMENTS</head><note type="other">Domain Dataset Method Test Accqm Accex Single WikiSQL</note><p>Seq2SQL <ref type="bibr" target="#b27">[28]</ref> 51.6% 60.4% SQLNet <ref type="bibr" target="#b23">[24]</ref> 61.3% 68.0% TypeSQL <ref type="bibr" target="#b25">[26]</ref>  </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.1">Evaluation</head><p>We conduct our preliminary experiments using a seq2seq model with stacked GRU. We use query-match accuracy Accqm for evaluation, we match synthesized queries against the ground truth p. We also compare the execution results as <ref type="bibr" target="#b27">[28]</ref>, denoted as Accex, if applicable.</p><p>We jointly train our multi-domain model on WikiSQL <ref type="bibr" target="#b27">[28]</ref>, Geo880 <ref type="bibr" target="#b26">[27]</ref> , and OVERNIGHT <ref type="bibr" target="#b22">[23]</ref>, their query types are SQL, Lambda expression, and SQL (we use the dataset that manually converted to SQL in <ref type="bibr" target="#b21">[22]</ref>). Some of the domains are over sampled to balance the number of training samples among all the domains. Our method is shown in Table <ref type="table" target="#tab_1">1</ref> as Ours-multi. Since all the domains are trained with in a single model, the accuracy of multi-domain model does not exhibit a large improvement. However, we believe it is a model capacity issue since the accuries of all the domains are very close or better than state-of-the-art performance. We observe that the seq2seq model is able to infer both SQL and Lambda expression as long as a tag (e.g., hSQLi, hlambdai ) indicating the query type is provided.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.2">Spatial Domain</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Domain Geo880 Restaurant Method</head><p>Accdm Method Accdm Single SEQ2TREE <ref type="bibr" target="#b1">[2]</ref> 87.1% PEK03 <ref type="bibr" target="#b17">[18]</ref> 97.0% TRANX <ref type="bibr" target="#b24">[25]</ref> 88.2% TM00 <ref type="bibr" target="#b20">[21]</ref> 99.6% JL16 <ref type="bibr" target="#b8">[9]</ref> 89.3% FKZ18 <ref type="bibr" target="#b2">[3]</ref> 100%</p><p>Ours-multi 90.7% 100%</p><p>Table <ref type="table">2</ref>: Comparison of models in Spatial Domain.</p><p>We conduct preliminary evaluations on the spatial domain (Geo880 <ref type="bibr" target="#b26">[27]</ref> and Restaurant <ref type="bibr" target="#b20">[21]</ref>), which is a more di cult task since the dataset in the spatial domain is sparse. We adopt the data element detector as a spatial comprehension model, and inject spatial semantics using symbol insertions. In this setting, we jointly train a multi-domain model with both Geo880 and Restaurant training sets (Restaurant is oversampled to balance all the domains), and evaluate on the test set of each separately (since both are for lambda expression queries, a prefix symbol is not inserted). As shown in Table <ref type="table">2</ref> (we use denotation match accuracy Acc dm for evaluation), our method (Ours-multi) outperforms previous methods.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>&lt;lambda&gt;Figure 2 :</head><label>2</label><figDesc>Figure 2: An example of cross-domain framework corresponds to Figure 1, di↵erent query types are treated equally.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 1 :</head><label>1</label><figDesc>Comparison of models.</figDesc><table><row><cell>75.4% 82.6%</cell></row></table></figure>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Natural language interfaces to databases-an introduction</title>
		<author>
			<persName><forename type="first">I</forename><surname>Androutsopoulos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">D</forename><surname>Ritchie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Thanisch</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Natural language engineering</title>
		<imprint>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="issue">01</biblScope>
			<biblScope unit="page" from="29" to="81" />
			<date type="published" when="1995">1995</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<monogr>
		<title level="m" type="main">Language to logical form with neural attention</title>
		<author>
			<persName><forename type="first">L</forename><surname>Dong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Lapata</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1601.01280</idno>
		<imprint>
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b2">
	<monogr>
		<title level="m" type="main">Improving text-to-sql evaluation methodology</title>
		<author>
			<persName><forename type="first">C</forename><surname>Finegan-Dollak</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">K</forename><surname>Kummerfeld</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Ramanathan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Sadasivam</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Radev</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1806.09029</idno>
		<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b3">
	<monogr>
		<title level="m" type="main">Adversarial and clean data are not twins</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Gong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W.-S</forename><surname>Ku</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1704.04960</idno>
		<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b4">
	<monogr>
		<title level="m" type="main">Adversarial texts with gradient methods</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Gong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Song</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W.-S</forename><surname>Ku</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1801.07175</idno>
		<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b5">
	<monogr>
		<title level="m" type="main">Explaining and harnessing adversarial examples</title>
		<author>
			<persName><forename type="first">I</forename><forename type="middle">J</forename><surname>Goodfellow</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Shlens</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Szegedy</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1412.6572</idno>
		<imprint>
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b6">
	<monogr>
		<title level="m" type="main">Neural semantic parsing over multiple knowledge-bases</title>
		<author>
			<persName><forename type="first">J</forename><surname>Herzig</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Berant</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1702.01569</idno>
		<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Learning a neural semantic parser from user feedback</title>
		<author>
			<persName><forename type="first">S</forename><surname>Iyer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Konstas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Cheung</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Krishnamurthy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Zettlemoyer</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">ACL</title>
		<imprint>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page" from="963" to="973" />
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<monogr>
		<title level="m" type="main">Data recombination for neural semantic parsing</title>
		<author>
			<persName><forename type="first">R</forename><surname>Jia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Liang</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<monogr>
		<title level="m" type="main">Adversarial examples for evaluating reading comprehension systems</title>
		<author>
			<persName><forename type="first">R</forename><surname>Jia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Liang</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1707.07328</idno>
		<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Googles multilingual neural machine translation system: Enabling zero-shot translation</title>
		<author>
			<persName><forename type="first">M</forename><surname>Johnson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Schuster</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><forename type="middle">V</forename><surname>Le</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Krikun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Thorat</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Viégas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Wattenberg</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Corrado</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">ACL</title>
		<imprint>
			<biblScope unit="volume">5</biblScope>
			<biblScope unit="page" from="339" to="351" />
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<monogr>
		<title level="m" type="main">Deep text classification can be fooled</title>
		<author>
			<persName><forename type="first">B</forename><surname>Liang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Su</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Bian</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Shi</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1704.08006</idno>
		<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Deepfool: a simple and accurate method to fool deep neural networks</title>
		<author>
			<persName><forename type="first">S.-M</forename><surname>Moosavi-Dezfooli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Fawzi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Frossard</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">CVPR</title>
				<imprint>
			<date type="published" when="2016">2016</date>
			<biblScope unit="page" from="2574" to="2582" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">The application of data mining techniques in financial fraud detection: A classification framework and an academic review of literature</title>
		<author>
			<persName><forename type="first">E</forename><surname>Ngai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Hu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Wong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Sun</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Decision Support Systems</title>
		<imprint>
			<biblScope unit="volume">50</biblScope>
			<biblScope unit="issue">3</biblScope>
			<biblScope unit="page" from="559" to="569" />
			<date type="published" when="2011">2011</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Application of data mining techniques in customer relationship management: A literature review and classification</title>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">W</forename><surname>Ngai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Xiu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">C</forename><surname>Chau</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Expert systems with applications</title>
		<imprint>
			<biblScope unit="volume">36</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page" from="2592" to="2602" />
			<date type="published" when="2009">2009</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">The limitations of deep learning in adversarial settings</title>
		<author>
			<persName><forename type="first">N</forename><surname>Papernot</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Mcdaniel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Jha</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Fredrikson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><forename type="middle">B</forename><surname>Celik</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Swami</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">IEEE EuroS&amp;P</title>
				<imprint>
			<publisher>IEEE</publisher>
			<date type="published" when="2016">2016. 2016</date>
			<biblScope unit="page" from="372" to="387" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Compositional semantic parsing on semi-structured tables</title>
		<author>
			<persName><forename type="first">P</forename><surname>Pasupat</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Liang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">ACL</title>
				<imprint>
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Towards a theory of natural language interfaces to databases</title>
		<author>
			<persName><forename type="first">A</forename><surname>Popescu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Etzioni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><forename type="middle">A</forename><surname>Kautz</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 8th International Conference on Intelligent User Interfaces</title>
				<meeting>the 8th International Conference on Intelligent User Interfaces</meeting>
		<imprint>
			<date type="published" when="2003">2003</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<monogr>
		<title level="m" type="main">Towards crafting text adversarial samples</title>
		<author>
			<persName><forename type="first">S</forename><surname>Samanta</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Mehta</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1707.02812</idno>
		<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b19">
	<monogr>
		<title level="m" type="main">Cross-domain semantic parsing via paraphrasing</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Su</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Yan</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1704.05974</idno>
		<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">Automated construction of database interfaces: Integrating statistical and relational learning for semantic parsing</title>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">R</forename><surname>Tang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">J</forename><surname>Mooney</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">ACL</title>
				<imprint>
			<date type="published" when="2000">2000</date>
			<biblScope unit="page" from="133" to="141" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<monogr>
		<title level="m" type="main">A transfer-learnable natural language interface for databases</title>
		<author>
			<persName><forename type="first">W</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Tian</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Xiong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W.-S</forename><surname>Ku</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1809.02649</idno>
		<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">Building a semantic parser overnight</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Berant</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Liang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">ACL</title>
		<imprint>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page" from="1332" to="1342" />
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<monogr>
		<title level="m" type="main">Sqlnet: Generating structured queries from natural language without reinforcement learning</title>
		<author>
			<persName><forename type="first">X</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Song</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1711.04436</idno>
		<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b24">
	<monogr>
		<title level="m" type="main">Tranx: A transition-based neural abstract syntax parser for semantic parsing and code generation</title>
		<author>
			<persName><forename type="first">P</forename><surname>Yin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Neubig</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1810.02720</idno>
		<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b25">
	<monogr>
		<title level="m" type="main">Typesql: Knowledge-based type-aware neural text-to-sql generation</title>
		<author>
			<persName><forename type="first">T</forename><surname>Yu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Radev</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1804.09769</idno>
		<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b26">
	<analytic>
		<title level="a" type="main">Learning to parse database queries using inductive logic programming</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">M</forename><surname>Zelle</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">J</forename><surname>Mooney</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">AAAI</title>
				<imprint>
			<date type="published" when="1996">1996</date>
			<biblScope unit="page" from="1050" to="1055" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b27">
	<monogr>
		<title level="m" type="main">Seq2sql: Generating structured queries from natural language using reinforcement learning</title>
		<author>
			<persName><forename type="first">V</forename><surname>Zhong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Xiong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Socher</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1709.00103</idno>
		<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
