<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>A cross-domain natural language interface to databases using adversarial text method</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Wenlu Wang</string-name>
          <email>wenluwang@auburn.edu</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>supervised by Wei-Shinn Ku</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Haixun Wang</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Auburn University</institution>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>and WeWork Research</institution>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2019</year>
      </pub-date>
      <abstract>
        <p>A natural language interface (NLI) to databases is an interface that supports natural language queries to be executed by database management systems (DBMS). However, most NLIs are domain specific due to the complexity of the natural language questions, and an NLI trained on one domain is hard to be transferred another due to the discrepancies between di↵erent ontology. Inspired by the idea of stripping domain-specific information out of natural language questions, we propose a cross-domain NLI with a general purpose question tagging strategy and a multi-language neural translation model. Our question tagging strategy is able to extract the “skeleton” of the question that represents its semantic structure for any domain. With question tagging, every domain will be handled equally with a single multi-language neural translation model. Our preliminary experiments show that our multi-domain model has excellent cross-domain transferability.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>INTRODUCTION</title>
      <p>
        Relational databases are widely adopted in real-world
applications [
        <xref ref-type="bibr" rid="ref14 ref15">15, 14</xref>
        ]. However, it requires a certain knowledge
of query languages to operate on DBMSs, which motivated
the study of NLI to databases [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] with the purpose of making
DBMSs operable by anyone without training.
      </p>
      <p>
        The challenges of NLI to databases lies in the
discrepancies of di↵erent ontology, which makes general-purpose NLI
hard to achieve. Most existing general purpose NLIs
exploit syntax-guided decoding and require the grammar of
the structured queries (domain specific grammar) as part of
the model. Such a model cannot be shared between di↵erent
grammars, while we propose a general purpose model where
di↵erent types of queries and di↵erence domains share the
same components. To overcome the obstacles of
generalizing one NLI model to di↵erent domains or even unseen
domains, we perform a pre-processing step inspired by the
strategy of separating domain-specific information from the
question [
        <xref ref-type="bibr" rid="ref22">22</xref>
        ]. By detaching domain-specific data elements,
the NLI model is able to focus on the semantic meaning and
agnostic of the natural language question, which facilitates
the cross-domain generalization.
      </p>
      <p>We first “strip” a natural language question (shown in
Figure 1), each type of the query (SQL and Lambda
expression in our examples) is treated equally, then translate
the tagged question to a structured query. The definition of
“strip” is enclosing a phrase that describes a data element
(tables, columns, values, keywords, etc.) appearing in the
query by inserting a symbol (k1, v1, etc.) representing the
type and index of the data element in front of the phrase,
and an “end of element” symbol (e.g., heoei) at the end
of the phrase. In Figure 1, we show two types of queries
(Lambda Expression and SQL), k represents a column field,
a table name, or a keyword, and v represents a value.</p>
      <p>Question
Query
Question
Query
Question
Query
Question
Query</p>
      <p>Which cities are located in Virginia ?
city(A), location(A, B), const(B, stateid(“Virginia”))
hlambdai Which hk1i cities heoei are hk2i located in heoei
hv2i Virginia heoei ?
k1(A), k2(A, B), const(B, stateid(v2))</p>
      <p>(1)
Which movies were scheduled to release on May 19 2019 ?
SELECT movie WHERE release date = May 19 2019
hSQLi Which hk1i movies heoei were scheduled
hk2i to release on heoei hv2i May 19 2019 heoei ?
SELECT k1 WHERE k2 = v2
(2)</p>
      <p>
        Another challenge of our cross-domain task is to handle
di↵erent query types. The aforementioned symbol insertion
strategy is able to handle questions of di↵erent types equally
but fails to di↵erentiate them. Inspired by Google’s
multilingual translation model [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] where an artificial token is
introduced at the beginning of the input sentence to
indicate the target language. We prefix a query type symbol to
indicate the target query type the NLI model should covert
to (e.g., hSQLi, hlambdai). For instance, consider the
following question –&gt; SQL pair:
      </p>
      <p>Which is the highest score? –&gt; SELECT MAX(score)
It will be modified to:</p>
      <p>hSQLi Which is the highest score? –&gt; SELECT MAX(score)
Such an approach only needs to prefix one additional token,
we will validate in the preliminary experiments that it is the
simplest but e↵ective approach.</p>
      <p>The core design of our symbol insertion strategy lies in
how to identify the phrase that describes a data element
appears in the corresponding query. The phrase might not
be the exact words of the data element. In Figure 1(1),
the data element “release date” is described as “to release
on” in the question. Inspired by gradient-based adversarial
text method, we propose an adversarial method towards a
data element detector. Given a natural language question q
and a data element e, the data element detector will predict
whether e is mentioned in q.</p>
      <p>Figure 1 presents two examples. In example (1), the
question (imply natural language question in this paper) is
converted to a lambda expression, and example (2) is converted
to a SQL query.
2.</p>
    </sec>
    <sec id="sec-2">
      <title>ADVERSARIAL TEXT METHOD</title>
      <p>
        It has been demonstrated that adding a carefully crafted
small noise is able to fool the deep neural network models
into wrong predictions, while the small noise makes
unnoticeable visual di↵erence [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. Most of the adversarial attack
methods on the text [
        <xref ref-type="bibr" rid="ref10 ref12 ref19">12, 19, 10</xref>
        ] try to perturb the features
(e.g., words, characters, and phrases) that are the most
influential on the predictions. Inspired by gradient-based
adversarial text attacks [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], we propose our own solution to
identify the position of a data element in a question.
3.
3.1
      </p>
    </sec>
    <sec id="sec-3">
      <title>DESIGN</title>
    </sec>
    <sec id="sec-4">
      <title>Overview</title>
      <p>
        Given a (question, query) pair, our core methodology is
to insert pre-designed symbols and enclose data elements
mentioned in the question to achieve the purpose of handling
every sample (of di↵erent domain/type) equally. Figure 2
shows the framework of our approach bottom-up.
1. Build a binary classifier BC as a data element
detector to predict whether a data element e appears in
a question q’s corresponding query p by the semantic
meaning of the question, which takes q and e as inputs
without referencing p.
2. Inspired by [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], we search for the most influential phrase
in the question using gradient-based adversarial text
methods. We refer “the most influential phrase” as
the phrase that describes the data element e
theoretically.
3. We insert symbols in q to enclose the phrases that
describes the data elements, denoted as q0. Since a
query type symbol (e.g., hSQLi) is prefixed to q0, a
user is able to select a desired query type.
4. Build a multi-lingual cross-domain sequence-to-sequence
(seq2seq) model to translate q0 to p0, and p0 is a query
where the data elements are replaced by symbols
inserted in q.
5. Inserted symbols are replaced with data elements to
form the original query (convert p0 back to p).
trained, it will produce positive predictions for data element
e = [“location”,“release date”]. As the true label, p is not
involved in the process. After translating q0 to p0, where
data elements are represented as symbols, we perform the
final step of converting p0 back to p.
      </p>
      <p>Data elements include table names, column fields, and
column values in databases, and keywords in query grammar.
For example, “movies”, “release date”, and “May 19 2019”
are data elements in SQL query “SELECT movie WHERE
release date EQUAL May 19 2019”, and the other
elements SELECT WHERE EQUAL belong to the template of
SQL sketch. In “city(A),location(A,B),const(B,
stateid(Virginia))”, “city” is a table name, “stateid”
is a column field, “Virginia” is a column value, and
“location” is a keyword in lambda expression grammar. The
challenge is to discover all the data elements from the
question , and for symbol insertion purpose, we need to discover
which phrase describes each data element. With such
symbols insertion, we relieve the burden of identify the data
elements from seq2seq model, and makes it focus on
learning semantic structure of the question and logic of the data
elements.</p>
      <p>We have two challenges to tackle for data element search
(use Figure 2 as an example):
1. Identify whether a data element is described in the
question. Ultimately, we are trying to detect all the
data elements that constitute the query, and we have
to infer those data elements from the natural language
question based on its semantics. In q1, we need to
identify all the data elements that are described in the
question (e.g., “city”, “location”, and “Virginia”). In
q2, we need to identify “release data”, “movie”, and
“May 19 2019”.
2. The phrase that describes a data element needs to be
identified by its semantic meaning and contextual
comprehension. In q1 of Figure 2, “located in” is identified
as the most influential phrase while describing the key
word “location”. In q2, “to release on” is identified as
the phrase that describes “release date”.</p>
      <p>To address these two challenges, we propose our general
purpose data element search strategy with two steps:
• We propose a Data Element Detector (Sec 3.2.1)
for the first challenge, which is a binary classifier with
a question q and data element e as an input. The
detector is able to detect whether the data element e is
mentioned in question q. As presented in Figure 2,
a Data Element Detector is shared among all the
domains.
• In the case of positive prediction in the previous step,
we propose an Adversarial Text Method (Sec 3.2.2)
for the second challenge, which relies on the
information learned by the binary classifier from the first step.</p>
      <sec id="sec-4-1">
        <title>3.2.1 Data Element Detector</title>
        <p>We use a bi-directional attentive recurrent neural network
to achieve question understanding. For a question q
composed of n tokens [q1, ..., qn] and a data element e composed
of m tokens [e1, ..., em], we use a pre-trained Glove
embedding to initiate a word embedding layer. On top of the
embedding layer, we use LSTM cells to produce hidden states
for each time step (each word in q). We build a similar
structure for e. We denote the top layer hidden states as
hq = [hq1, · · · , hqn]
he = [he1, · · · , hem]
We build a bi-directional LSTM layer on hq with attention
over he.
where W0, W1, W2 ,and v are model parameters. Here t
enumerates each time step for e, and j enumerates each
token in q. We compute bi-directional output dt =! [dt , dt ],
and feed it to a multi-layer perception for binary prediction.
3.2.2</p>
      </sec>
      <sec id="sec-4-2">
        <title>Adversarial Text Method</title>
        <p>
          With the adversarial text method, given a data element e
that has a positive prediction from the binary classifier, we
propose to search a phrase of the question that describes e.
We describe our adversarial text method as follows.
1. We have trained a Data Element Detector that takes a
question q and a data element e as inputs and predicts
whether e is described in q.
2. We search for the most influential phrase in q using
gradient-based adversarial text methods [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ]. There are
three possible directions (the loss gradient of the Data
Element Detector with q and e as inputs is r L(q, e)):
- DeepFool [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ]. We iteratively search the optimal
direction in which only a minimum perturbation is
needed to a↵ect the prediction. Theoretically, the
optimal direction is ||r fL((qq,,ee))||22 r L(q, e) where f (·)
denotes the Data Element Detector.
- Fast Gradient Method FGM [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ]. We add a noise
that is proportional to either r L(q, e) or sign(r L(q, e))
to the original sample to change the prediction of the
Test
Accqm
51.6%
61.3%
75.4%
74.5%
76.8%
84.1%
        </p>
        <p>
          Accex
60.4%
68.0%
82.6%
82.7%
Data Element Detector. In particular, the noise for
each token qi is proportional to @L (q,e) .
@q i
- JSMA [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ]. We calculate the Jacobian-based saliency
map based on r L(q, e), and perturb one token at a
time. The chosen token has the highest saliency value.
Since all the methods are trying to add minimum noise
that influences the prediction the most. The locations
where the noise is added will be the positions of tokens
that constitute the most influential phrase – i,e., the
phrase that describes the data element e.
3. We search for a phrase in the question where adding a
small perturbation will a↵ect the prediction
dramatically.
        </p>
        <p>The challenge of our adversarial text method is the
discreteness of the text domain. Words or characters are discrete
variables thus indi↵erentiable. To overcome such problem,
we propose to calculate the loss gradient (r L) of the
target model w.r.t. the embedding of each word, where the
embedding space is smooth.
3.3</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Neural Machine Translation</title>
      <p>We denote an question post symbol insertion as q0 and
the corresponding query post symbol replacement as p0. We
train a seq2seq model to translate q0 to p0:</p>
      <p>p0 = seq2seq(q0)
Encoder is a stacked bi-directional multi-layer recurrent
neural network (RNN). Decoder is a one-layer attentive RNN.</p>
      <p>We use a single multilingual neural translation model for
our cross-domain NLI task. We believe with a prefixed query
type symbol (e.g., hSQLi), a multi-domain model is able to
handle di↵erent query types, and each query type is treated
equally.
4.</p>
    </sec>
    <sec id="sec-6">
      <title>RELATED WORK</title>
      <p>
        NLI to databases was first formally introduced in [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ].
Semantic parsing [
        <xref ref-type="bibr" rid="ref17 ref23 ref9">17, 23, 9</xref>
        ] and cross-domain semantic
parsing [
        <xref ref-type="bibr" rid="ref20 ref7">7, 20</xref>
        ] are applied in NLI to databases. However, due to
the incompatibility among di↵erent domains, cross-domain
task remains unsolved. The sketch-based solutions are also
extensively studied, which is first proposed in [
        <xref ref-type="bibr" rid="ref24">24</xref>
        ]. A deep
model is trained to fill the slots in the sketch. An
extension of sketch-based solution [
        <xref ref-type="bibr" rid="ref26">26</xref>
        ] relies on a knowledge base
to identify the column values. Such a strategy is confined
in pre-defined sketch and existing knowledge base. seq2seq
model are also exploited to serve as a translator [
        <xref ref-type="bibr" rid="ref28 ref8">8, 28</xref>
        ],
which has no limitations on query sketch.
5.
      </p>
    </sec>
    <sec id="sec-7">
      <title>PRELIMINARY EXPERIMENTS</title>
      <p>Domain
Dataset
Method
Single
Multi
WikiSQL</p>
      <p>
        WikiSQL
OVERNIGHT
Geo880
Seq2SQL [
        <xref ref-type="bibr" rid="ref28">28</xref>
        ]
SQLNet [
        <xref ref-type="bibr" rid="ref24">24</xref>
        ]
TypeSQL [
        <xref ref-type="bibr" rid="ref26">26</xref>
        ]
      </p>
    </sec>
    <sec id="sec-8">
      <title>Evaluation</title>
      <p>
        We conduct our preliminary experiments using a seq2seq
model with stacked GRU. We use query-match accuracy
Accqm for evaluation, we match synthesized queries against
the ground truth p. We also compare the execution results
as [
        <xref ref-type="bibr" rid="ref28">28</xref>
        ], denoted as Accex, if applicable.
      </p>
      <p>
        We jointly train our multi-domain model on WikiSQL [
        <xref ref-type="bibr" rid="ref28">28</xref>
        ],
Geo880 [
        <xref ref-type="bibr" rid="ref27">27</xref>
        ] , and OVERNIGHT [
        <xref ref-type="bibr" rid="ref23">23</xref>
        ], their query types are
SQL, Lambda expression, and SQL (we use the dataset that
manually converted to SQL in [
        <xref ref-type="bibr" rid="ref22">22</xref>
        ]). Some of the domains
are over sampled to balance the number of training samples
among all the domains. Our method is shown in Table 1
as Ours-multi. Since all the domains are trained with in
a single model, the accuracy of multi-domain model does
not exhibit a large improvement. However, we believe it is a
model capacity issue since the accuries of all the domains are
very close or better than state-of-the-art performance. We
observe that the seq2seq model is able to infer both SQL and
Lambda expression as long as a tag (e.g., hSQLi, hlambdai
) indicating the query type is provided.
5.2
      </p>
    </sec>
    <sec id="sec-9">
      <title>Spatial Domain</title>
      <p>
        We conduct preliminary evaluations on the spatial domain
(Geo880 [
        <xref ref-type="bibr" rid="ref27">27</xref>
        ] and Restaurant [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ]), which is a more dicult
task since the dataset in the spatial domain is sparse. We
adopt the data element detector as a spatial comprehension
model, and inject spatial semantics using symbol insertions.
In this setting, we jointly train a multi-domain model with
both Geo880 and Restaurant training sets (Restaurant is
oversampled to balance all the domains), and evaluate on
the test set of each separately (since both are for lambda
expression queries, a prefix symbol is not inserted). As shown
in Table 2 (we use denotation match accuracy Accdm for
evaluation), our method (Ours-multi ) outperforms previous
methods.
      </p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>I.</given-names>
            <surname>Androutsopoulos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G. D.</given-names>
            <surname>Ritchie</surname>
          </string-name>
          , and
          <string-name>
            <given-names>P.</given-names>
            <surname>Thanisch</surname>
          </string-name>
          .
          <article-title>Natural language interfaces to databases-an introduction</article-title>
          .
          <source>Natural language engineering</source>
          ,
          <volume>1</volume>
          (
          <issue>01</issue>
          ):
          <fpage>29</fpage>
          -
          <lpage>81</lpage>
          ,
          <year>1995</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>L.</given-names>
            <surname>Dong</surname>
          </string-name>
          and
          <string-name>
            <given-names>M.</given-names>
            <surname>Lapata</surname>
          </string-name>
          .
          <article-title>Language to logical form with neural attention</article-title>
          .
          <source>arXiv preprint arXiv:1601.01280</source>
          ,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>C.</given-names>
            <surname>Finegan-Dollak</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. K.</given-names>
            <surname>Kummerfeld</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Ramanathan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Sadasivam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , and
          <string-name>
            <given-names>D.</given-names>
            <surname>Radev</surname>
          </string-name>
          .
          <article-title>Improving text-to-sql evaluation methodology</article-title>
          .
          <source>arXiv preprint arXiv:1806.09029</source>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Gong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Wang</surname>
          </string-name>
          , and
          <string-name>
            <given-names>W.-S.</given-names>
            <surname>Ku</surname>
          </string-name>
          .
          <article-title>Adversarial and clean data are not twins</article-title>
          .
          <source>arXiv preprint arXiv:1704.04960</source>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Gong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Song</surname>
          </string-name>
          , and
          <string-name>
            <given-names>W.-S.</given-names>
            <surname>Ku</surname>
          </string-name>
          .
          <article-title>Adversarial texts with gradient methods</article-title>
          . arXiv preprint arXiv:
          <year>1801</year>
          .07175,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>I. J.</given-names>
            <surname>Goodfellow</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Shlens</surname>
          </string-name>
          , and
          <string-name>
            <given-names>C.</given-names>
            <surname>Szegedy</surname>
          </string-name>
          .
          <article-title>Explaining and harnessing adversarial examples</article-title>
          .
          <source>arXiv preprint arXiv:1412.6572</source>
          ,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>J.</given-names>
            <surname>Herzig</surname>
          </string-name>
          and
          <string-name>
            <given-names>J.</given-names>
            <surname>Berant</surname>
          </string-name>
          .
          <article-title>Neural semantic parsing over multiple knowledge-bases</article-title>
          .
          <source>arXiv preprint arXiv:1702.01569</source>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>S.</given-names>
            <surname>Iyer</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Konstas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Cheung</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Krishnamurthy</surname>
          </string-name>
          , and
          <string-name>
            <given-names>L.</given-names>
            <surname>Zettlemoyer</surname>
          </string-name>
          .
          <article-title>Learning a neural semantic parser from user feedback</article-title>
          .
          <source>In ACL</source>
          , volume
          <volume>1</volume>
          , pages
          <fpage>963</fpage>
          -
          <lpage>973</lpage>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>R.</given-names>
            <surname>Jia</surname>
          </string-name>
          and
          <string-name>
            <given-names>P.</given-names>
            <surname>Liang</surname>
          </string-name>
          .
          <article-title>Data recombination for neural semantic parsing</article-title>
          .
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>R.</given-names>
            <surname>Jia</surname>
          </string-name>
          and
          <string-name>
            <given-names>P.</given-names>
            <surname>Liang</surname>
          </string-name>
          .
          <article-title>Adversarial examples for evaluating reading comprehension systems</article-title>
          .
          <source>arXiv preprint arXiv:1707.07328</source>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>M.</given-names>
            <surname>Johnson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Schuster</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q. V.</given-names>
            <surname>Le</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Krikun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Thorat</surname>
          </string-name>
          , F. Vi´egas, M. Wattenberg,
          <string-name>
            <given-names>G.</given-names>
            <surname>Corrado</surname>
          </string-name>
          , et al.
          <article-title>Googles multilingual neural machine translation system: Enabling zero-shot translation</article-title>
          .
          <source>ACL</source>
          ,
          <volume>5</volume>
          :
          <fpage>339</fpage>
          -
          <lpage>351</lpage>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>B.</given-names>
            <surname>Liang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Su</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Bian</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Li</surname>
          </string-name>
          , and
          <string-name>
            <given-names>W.</given-names>
            <surname>Shi</surname>
          </string-name>
          .
          <article-title>Deep text classification can be fooled</article-title>
          .
          <source>arXiv preprint arXiv:1704.08006</source>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>S.-M.</given-names>
            <surname>Moosavi-Dezfooli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Fawzi</surname>
          </string-name>
          , and
          <string-name>
            <given-names>P.</given-names>
            <surname>Frossard</surname>
          </string-name>
          .
          <article-title>Deepfool: a simple and accurate method to fool deep neural networks</article-title>
          .
          <source>In CVPR</source>
          , pages
          <fpage>2574</fpage>
          -
          <lpage>2582</lpage>
          ,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>E.</given-names>
            <surname>Ngai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Hu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Chen</surname>
          </string-name>
          , and
          <string-name>
            <surname>X. Sun.</surname>
          </string-name>
          <article-title>The application of data mining techniques in financial fraud detection: A classification framework and an academic review of literature</article-title>
          .
          <source>Decision Support Systems</source>
          ,
          <volume>50</volume>
          (
          <issue>3</issue>
          ):
          <fpage>559</fpage>
          -
          <lpage>569</lpage>
          ,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>E. W.</given-names>
            <surname>Ngai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Xiu</surname>
          </string-name>
          , and
          <string-name>
            <given-names>D. C.</given-names>
            <surname>Chau</surname>
          </string-name>
          .
          <article-title>Application of data mining techniques in customer relationship management: A literature review and classification</article-title>
          .
          <source>Expert systems with applications</source>
          ,
          <volume>36</volume>
          (
          <issue>2</issue>
          ):
          <fpage>2592</fpage>
          -
          <lpage>2602</lpage>
          ,
          <year>2009</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>N.</given-names>
            <surname>Papernot</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>McDaniel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Jha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Fredrikson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z. B.</given-names>
            <surname>Celik</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <given-names>A.</given-names>
            <surname>Swami</surname>
          </string-name>
          .
          <article-title>The limitations of deep learning in adversarial settings</article-title>
          .
          <source>In 2016 IEEE EuroS&amp;P</source>
          , pages
          <fpage>372</fpage>
          -
          <lpage>387</lpage>
          . IEEE,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>P.</given-names>
            <surname>Pasupat</surname>
          </string-name>
          and
          <string-name>
            <given-names>P.</given-names>
            <surname>Liang</surname>
          </string-name>
          .
          <article-title>Compositional semantic parsing on semi-structured tables</article-title>
          .
          <source>In ACL</source>
          ,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>A.</given-names>
            <surname>Popescu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Etzioni</surname>
          </string-name>
          , and
          <string-name>
            <given-names>H. A.</given-names>
            <surname>Kautz</surname>
          </string-name>
          .
          <article-title>Towards a theory of natural language interfaces to databases</article-title>
          .
          <source>In Proceedings of the 8th International Conference on Intelligent User Interfaces</source>
          ,
          <year>2003</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>S.</given-names>
            <surname>Samanta</surname>
          </string-name>
          and
          <string-name>
            <given-names>S.</given-names>
            <surname>Mehta</surname>
          </string-name>
          .
          <article-title>Towards crafting text adversarial samples</article-title>
          .
          <source>arXiv preprint arXiv:1707.02812</source>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Su</surname>
          </string-name>
          and
          <string-name>
            <given-names>X.</given-names>
            <surname>Yan</surname>
          </string-name>
          .
          <article-title>Cross-domain semantic parsing via paraphrasing</article-title>
          .
          <source>arXiv preprint arXiv:1704.05974</source>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>L. R.</given-names>
            <surname>Tang</surname>
          </string-name>
          and
          <string-name>
            <given-names>R. J.</given-names>
            <surname>Mooney</surname>
          </string-name>
          .
          <article-title>Automated construction of database interfaces: Integrating statistical and relational learning for semantic parsing</article-title>
          .
          <source>In ACL</source>
          , pages
          <fpage>133</fpage>
          -
          <lpage>141</lpage>
          ,
          <year>2000</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>W.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Tian</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Xiong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Wang</surname>
          </string-name>
          , and
          <string-name>
            <given-names>W.-S.</given-names>
            <surname>Ku</surname>
          </string-name>
          .
          <article-title>A transfer-learnable natural language interface for databases</article-title>
          .
          <source>arXiv preprint arXiv:1809.02649</source>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Berant</surname>
          </string-name>
          , and
          <string-name>
            <given-names>P.</given-names>
            <surname>Liang</surname>
          </string-name>
          .
          <article-title>Building a semantic parser overnight</article-title>
          .
          <source>In ACL</source>
          , volume
          <volume>1</volume>
          , pages
          <fpage>1332</fpage>
          -
          <lpage>1342</lpage>
          ,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>X.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Liu</surname>
          </string-name>
          , and
          <string-name>
            <given-names>D.</given-names>
            <surname>Song</surname>
          </string-name>
          . Sqlnet:
          <article-title>Generating structured queries from natural language without reinforcement learning</article-title>
          .
          <source>arXiv preprint arXiv:1711.04436</source>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>P.</given-names>
            <surname>Yin</surname>
          </string-name>
          and
          <string-name>
            <given-names>G.</given-names>
            <surname>Neubig. Tranx</surname>
          </string-name>
          :
          <article-title>A transition-based neural abstract syntax parser for semantic parsing and code generation</article-title>
          .
          <source>arXiv preprint arXiv:1810.02720</source>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [26]
          <string-name>
            <given-names>T.</given-names>
            <surname>Yu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , and
          <string-name>
            <given-names>D.</given-names>
            <surname>Radev</surname>
          </string-name>
          . Typesql:
          <article-title>Knowledge-based type-aware neural text-to-sql generation</article-title>
          .
          <source>arXiv preprint arXiv:1804.09769</source>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [27]
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Zelle</surname>
          </string-name>
          and
          <string-name>
            <given-names>R. J.</given-names>
            <surname>Mooney</surname>
          </string-name>
          .
          <article-title>Learning to parse database queries using inductive logic programming</article-title>
          .
          <source>In AAAI</source>
          , pages
          <fpage>1050</fpage>
          -
          <lpage>1055</lpage>
          ,
          <year>1996</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          [28]
          <string-name>
            <given-names>V.</given-names>
            <surname>Zhong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Xiong</surname>
          </string-name>
          , and
          <string-name>
            <given-names>R.</given-names>
            <surname>Socher</surname>
          </string-name>
          . Seq2sql:
          <article-title>Generating structured queries from natural language using reinforcement learning</article-title>
          .
          <source>arXiv preprint arXiv:1709.00103</source>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>