<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>"Gimme the Usual" - How Handling of Pragmatics Improves Chatbots</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Alessia Bianchini</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Francesco Tarasconi</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Raffaella Ventaglio</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Mariafrancesca Guadalupi CELI Language Technology</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>bianchini</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>tarasconi</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>ventaglio</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>guadalupi}@celi.it</string-name>
        </contrib>
      </contrib-group>
      <abstract>
        <p>English. We provide our view on the components needed for both the development and further improvement of robust and effective chatbots. We focus on why Pragmatics is important in developing next generation chatbots by bringing a few generalizable examples. We report our current experience on the design and implementation of a task-oriented textual chatbot for a closed-domain Question Answering system, which tackles problems in Pragmatics.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        Why Pragmatics Matters in Chatbots
Chatbot, chatterbot, natural-language interface,
dialogue system are some of the terms used to
refer to softwares that aim to carry on conversations
with humans
        <xref ref-type="bibr" rid="ref4">(Mauldin, 1994; Lester, Branting
and Mott, 2004; Boualem, Casati and Toumani,
2004)</xref>
        . We will not go into further details about
the classification and definition of such softwares.
We will use chatbot as if it was a hypernym of the
above mentioned softwares instead.
      </p>
      <p>
        Chatbots and Intent Understanding. The
goal of an intelligent chatbot is to understand the
user’s intent
        <xref ref-type="bibr" rid="ref15">(Yue, 2017)</xref>
        and behave accordingly.
Such goal is quite complex to achieve, and beyond
the capability of current state of the art chatbots.
However, the hype around chatbots has raised
awareness of what elements are needed for a
chatbot to manage human-like interactions. It is
generally agreed that to build effective and solid
chatbots the following is needed:
      </p>
      <p>Natural Language Processing (NLP): much
of the intelligence needed to understand human
intent lies in the processing of human language.
Hence, the development and improvement of
NLP algorithms is a necessary prerequisite for the
creation of intelligent chatbots.</p>
      <p>
        Machine Learning (ML): chatbot design should
rely on ML for learning and automatically
consolidating NLP rules by means of observation
of past experience - i.e., past conversations and
their outcomes
        <xref ref-type="bibr" rid="ref8">(Perez-Marin, 2011)</xref>
        . Current
chatbot development, given enough annotated
data, should consider adopting recently developed
algorithms that are task-oriented (Bordes and
Weston, 2016) or topic aware
        <xref ref-type="bibr" rid="ref14">(Xing, Chen et al.,
2017)</xref>
        . Developments in reinforcement learning
applications seem promising for task-oriented
dialogue systems
        <xref ref-type="bibr" rid="ref10 ref8">(Rieser and Lemon, 2011)</xref>
        .
      </p>
      <p>
        Context and State Awareness: depending on the
purpose of the chatbot, the component responsible
for the managing of the conversation (Dialogue
Manager System - DMS) should take into account
both context and states variables
        <xref ref-type="bibr" rid="ref2 ref6">(Allen, Byron,
Dzikovska, Ferguson, Galescu and Stent, 2001)</xref>
        .
From the DMS point of view, chatbots are usually
classified as: stateless chatbots; semi-stateful
chatbots; stateful chatbots (next generation
chatbots). During the conversation, state transitions
depend on the information acquired before. As for
the follow up action, it depends on the recognized
context.
      </p>
      <p>
        Natural Language Generation (NLG): NLG
concerns what information and in what form it
should be delivered (Breen, 2014). Dealing with
"real" conversation requires being both proactive
(e.g. suggest the best option; drive along the
compilation of a form; remind planned activities;
...)
        <xref ref-type="bibr" rid="ref6">(Owen et al., 2001)</xref>
        and adaptive (e.g. change
style - both in written and spoken scenarios
- according to domain, mood of the user, or
sociolinguistic variables).
      </p>
      <p>
        We argue here that, in addition to the above
mentioned, moving from Semantics to Pragmatics
plays a crucial role in building chatbots. This
is because a lot of the knowledge human beings
share during a conversation gets constructed along
the conversation itself
        <xref ref-type="bibr" rid="ref11 ref7">(Robyn, 2002; Pask, 1975)</xref>
        .
For instance, let’s consider the following mock
dialogue between Human (H) and Chatbot (C):
H lands on a money transfer service page.
H: Hello, I would like to make a transfer
C: Hello. Sure. Would you like to know more
about: [FAQ menu about transfer service is
shown]?
H navigates the FAQ menu
H signs up online to proceed with the transfer.
H: I would like to make a transfer
C: Sure. It only takes a couple of minutes
C starts the procedure to execute the transfer.
      </p>
      <p>In this interaction, the sentence “I would like
to make a transfer” instantiates two different
intents: informative intent, at first (H is looking
for information about the transfer service); follow
up intent, then (once H is satisfied with transfer
service conditions, he/she wants to proceed with
the transfer). Such pragmatic disambiguation
involves taking knowledge from the conversational
context into account, which is one of the most
difficult tasks for a chatbot. We report below how
we deal with this task in our task-oriented
closeddomain chatbot.
2 Intent Understanding in Practice
Understanding intents implies handling both
semantic meaning and pragmatic meaning. Roughly
speaking, while semantics concerns the meaning
of a sentence from the linguistic point of view,
Pragmatics concerns the interpretation of the same
sentence depending on extralinguistic knowledge
(Grice, 1975). As mentioned before, a sentence
can be ambiguous from the intent point of view.
As for classifying intents, there seems to be no
comprehensive literature about it, yet - not from
the chatbot perspective, at least. However, based
on our business experience, we would arrange
intents as follows:</p>
      <p>Informative Intent: the user is looking for
information; e.g. Question and Answer (QA),
FAQ browsing typically instantiate this intent.</p>
      <p>
        Follow Up Intent: as in regular conversations,
the user wants to "do things with words"
        <xref ref-type="bibr" rid="ref1">(Austin,
1962)</xref>
        , perform actions; e.g. "Call the call center",
"Order pizza", "Turn on washing machine".
      </p>
      <p>Dialogic Intent: the user uses discourse
markers to connect, organize, manage the conversation;
e.g.: greetings, farewells, turns markers, ...</p>
      <p>Regular expressions, pattern matching and
keyword recognition typically are not enough to
achieve real intent understanding. This is because
the more the interaction is human-like, the more
complicated it becomes to figure out what the
human really wants. Among business intent, real
life cases we faced are: Onboarding, Question
Answering and Education. In our applications,
we break down the understanding process into
subtasks. Namely: intent classification (e.g.
“booking a flight”); slot filling, i.e. enriching
the intent with more detailed information (such
as “destination” and “departure time”); context
modeling, i.e. keeping track of context to get to
the correct meaning (“time” might refer to “flight
departure time”, “flight arrival time”, “dinner
time”, etc...).</p>
    </sec>
    <sec id="sec-2">
      <title>System Design and Architecture</title>
      <p>
        Our task-oriented closed-domain financial textual
chatbot, Financial QA Chatbot, aims to provide
users with answers concerning banks and
insurances, through a conversation in Italian. The type
of answers that a user can obtain are similar to the
ones found on a financial platform website1: this
portal provides a search engine and FAQ section
to satisfy the information need. Therefore, it is
mainly a QA chatbot, although some additional
follow up actions are available on top of providing
an answer to questions, such as redirecting to
specific websites or services. Financial QA is
provided with a proprietary scoring algorithm to
match the current user’s questions to answers in
a database A. In line with previous work
        <xref ref-type="bibr" rid="ref9">(Quarteroni and Manandhar, 2007)</xref>
        , we will review key
design and architecture aspects, with emphasis
on possible solutions to Pragmatics problems
discussed in sections 1 and 2. In this sense, the
most significant components are the Dialogue
Manager and the Context Manager, which provide
the scoring algorithm enriched information.
NLP functions such as normalization,
tokenization, lemmatization, POS tagging, disambiguation
and dependency parsing are made available
through the CELI linguistic pipeline2
        <xref ref-type="bibr" rid="ref13">(Tarasconi
and Di Tomaso, 2015)</xref>
        .
      </p>
      <p>Dialogue Scenario: a QA session consists
of actions that can be performed by the user or
by the automated system, according to Dialogue
Management logic.</p>
      <p>User actions: greet, quit, ask a question q,
acknowledge the previous utterance, ask for
help/suggestions, browse the navigation menu.
System actions: greet, quit, present answer a,
acknowledge the previous utterance, ask for
clarifications, propose a follow up (question/action),
reprimand for using swearwords, suggest
questions, present or hide the navigation menu.</p>
      <p>
        User’s action classification: each user’s
utterance is classified into one of five action classes:
greet, quit, ask a question, acknowledge, ask
for help. This is accomplished using predefined
dictionaries and automatic classifiers, which also
consider discourse markers and disfluencies.
Although there is promising work done on dialog
1gooruf.com
2www.celi.it
act detection with multi-level information
        <xref ref-type="bibr" rid="ref12">(Rosset
et al., 2008)</xref>
        , in this step with adopt a simpler
approach, leaving further refinements to subsequent
components.
      </p>
      <p>Dialogue Management: the conversation
proceeds along these logics.</p>
      <p>1. An initial greeting (greet action), a request
for help (ask for help) or a direct question q
(ask a question) from the user.
2. The system, if asked for help, presents the
user with a navigation menu, based on
current context and on the given hierarchical
classification of contexts or topics (see
Context Management below). This menu can be
browsed until a terminal node in the
classification is reached, and, at that point, a
predefined set of questions related to that topic is
suggested. The user can select a question q
from that list.
3. q is analyzed to detect wh-type (Huang et al.,
2008) and whether it is elliptic or anaphoric.
This information is passed along with q and
the current context to the subsequent QA
component.
4. The QA component searches for matches of
the query according to the QA Algorithm.
Each matching answer ak is accompanied by
a relevance score rk, rk 2 (0; 1]. If at least
one match has relevance more than a fixed
threshold T , only the best match (highest
relevance) is returned. Otherwise, up to the top
Nr highest results are returned by the QA
component. In Financial QA’s basic settings,
T = 0:75 and Nr = 5.
5. The QA component results are processed:
they can be a single answer or, because of low
relevance scores, a list of answers. If a single
answer is provided by the QA component, it
is returned to the user (answer action). In the
case of a list, the user is asked for
clarifications, and a single answer is selected based
on her additional input (ask for clarifications
action, then answer action). After an answer
is provided to the user, context is updated
accordingly.
6. The system inquires whether the user is
interested in a follow up session; if this is the case,
the user can enter a question again. Else, the
system acknowledges.</p>
      <p>Ask me a question or choose one of the
following topics:
7. Whenever the user wants to terminate the
interaction, a final greeting is exchanged (quit
action).</p>
      <p>Context Management: intuitively, all the
answers a in the knowledge base are grouped in
disjoint topics of maximum granularity, which are
then organized in a hierarchical structure, used to
model context in this QA task.</p>
      <p>Managing topic hierarchies can improve
performance in a query matching system (Domingues
et al., 2014). Formally, context elements are
topics of conversation belonging to the finite set C =
fC1; : : : ; CN g. Topics are arranged in a
hierarchical classification structure, which can be
represented as a tree T = (C; E), where C is the set of
nodes. Edges E express the "Ci has subclass Cj "
relation. A context X is, in general, an arbitrarily
ordered sequence of topics.</p>
      <p>In our current implementation of Financial QA,
we support only contexts of length 1, therefore the
context Xs at step s of the conversation is the
position Cs in T. We assume all interactions start at
the root node C0. Xs is meant to represent the
current topic of conversation at step s, according
to the last answer provided or the latest click on
the navigation menu. By supporting contexts of
length &gt; 1, it is also possible to keep track of
previous topics of conversation.</p>
      <p>Each node Ci has a corresponding nonempty set
of topic-related keywords Wi.</p>
      <p>An important distinction is drawn between
terminal nodes C and nonterminal nodes C n C .
Each terminal node Cj has a (potentially empty)
set of answers Aj corresponding to it. All the Aj
sets are disjointed. Let A be the set of all answers:
A = [j21;:::;!Aj .</p>
      <p>In our current implementation, there are 35
classification nodes arranged on 3 levels, 25 of them
are terminal ones; the number of answers in the
knowledge base is 440, and growing
In the Financial QA chatbot, two types of moves
between contexts in C are allowed:
1. To children nodes or root node: using the
interactive navigation menu. Context is
updated automatically according to the user’s
selection.</p>
      <p>Example: You are in the “People” section.
(a) members
(b) influencers
(c) contact us
(d) return to main menu
2. To any terminal node: after answer ak is
provided to the user by the system, new
context becomes Cj , where Aj contains ak.
Example: after providing the answer
COST_OF_GOORUF = "Gooruf is free,
only Premium Providers are required to pay",
context is changed to ROOT ! services !
info_about_gooruf, info_about_gooruf
being the terminal topic containing answer
COST_OF_GOORUF.</p>
      <p>QA Algorithm Design: question q, its wh-type
h, and its current context Cs are passed to the
algorithm. Keywords wq are extracted from q. If
q is anaphoric or elliptic, the algorithm evaluates
whether to expand Wq to w_ q by using keywords
Ws corresponding to Cs. The final representation
of q is:</p>
      <p>R(q) = (Cs; h; w_ q):
Answers a 2 A are described by the following
feature vector:</p>
      <p>F (a) = (Ca ; Ha; Wa)
where Ca is the classification terminal node
corresponding to a, Wa the corresponding keywords
and Ha a set of related wh-type (for example a
user might inquire about Gooruf by referring to it
as a what or a who).</p>
      <p>Relevance rk for each answer ak is computed, by
considering the classification structure T as well,
therefore:</p>
      <p>
        rk(q) = (R(q); F (ak); T ):
To compute , scores are calculated separately
by comparing contexts (using proximity in T
between Cs and Ca in T ), wh-types (h and Ha)
and a dense semantic representation of keywords
(w_ q and Wa) obtained using a Word2Vec model
for Italian language
        <xref ref-type="bibr" rid="ref5">(Mikolov et al., 2013)</xref>
        ; before
these partial scores are weighted and summed.
      </p>
      <p>Example: we provide below an example of
Financial QA interaction which shows how
managing hierarchical context helps in accomplishing
the question answering task.</p>
      <p>A subtree taxes of T models Italian taxes-related
topics:
taxes ! city_taxes
taxes ! income_taxes
– TARI
– TASI
– IRPEF
– IRAP
Individual taxes are represented as terminal nodes
in T . Let Ctaxes = fIRPEF, IRAP, TARI, TASIg.
Each t 2 Ctaxes has associated answers:
"HOW_TO_PAY t", "WHERE_TO_PAY t",
"AMOUNT_TO_PAY t".</p>
      <p>The interaction could go as follows:
H: How can I pay city taxes?
QA Algorithm detects wh-type how, keywords
matching the city_taxes node, finds two relevant
answers in children nodes and Chatbot asks for
clarification.</p>
      <p>C: Did you mean TARI or TASI?
H: the second one
Chatbot presents answer "HOW_TO_PAY TASI"
Context is now TASI.</p>
      <p>H: and where can I pay it?
QA Algorithm detects wh-type where and
completes the question with context knowledge.
Chatbot finds a single relevant answer and
presents answer "WHERE_TO_PAY TASI".
H: how much IRPEF should I pay?
Chatbot presents "AMOUNT_TO_PAY IRPEF".
Context is now IRPEF.</p>
      <p>H: where can I pay it?
Chatbot presents "WHERE_TO_PAY IRPEF".
4</p>
    </sec>
    <sec id="sec-3">
      <title>Conclusions and Further Work</title>
      <p>
        We are currently in the process of evaluating
Financial QA according to a framework based
on PARADISE
        <xref ref-type="bibr" rid="ref10 ref8">(Walker et al., 1997; Rieser and
Lemon, 2011)</xref>
        , which considers, among the
others, the following indicators: Task Ease, NLU
Performance, Expected Behavior, Presentation,
Verbal Presentation, Future Use. We plan to finalize
our evaluation in the next months.
      </p>
      <p>NLP is crucial for the development of robust
chatbots; since extra-linguistic elements are
potentially very important in intent understanding,
moving from semantics to pragmatics is a necessary
step to develop next-generation chatbots. We have
shown how Dialogue Management can support a
more robust handling of context, at least in
closeddomain QA tasks.</p>
      <p>
        Further work is required to handle more business
cases and a broader definition of context, such as
history of activities conducted by the same user,
which can be especially useful in chatbots with
recommender functions
        <xref ref-type="bibr" rid="ref3">(Lombardi et al., 2009)</xref>
        .
We would like to thank Andrea Bolioli for his help
in the review phase and all of our CELI colleagues
for their invaluable work and support.
      </p>
      <p>Antoine Bordes and Jason</p>
      <p>End-to-End Goal-Oriented
arXiv:1605.07683 [cs.CL].</p>
      <p>Weston.</p>
      <p>Dialog.</p>
      <p>Learning
2016.</p>
      <p>Benatallah Boualem, Fabio Casati, and Farouk
Toumani. 2004. Web service conversation
modeling: A cornerstone for e-business automation. IEEE
Internet computing, 8(1):46–54.</p>
      <p>Andrew Breen 2014. Creating Expressive TTS Voices
for Conversation Agent Applications. International
Conference on Speech and Computer. Springer.
Marcos Aurélio Domingues, Marcelo Garcia Manzato,
Ricardo Marcondes Marcacini, Camila Vaccari
Sundermann and Solange Oliveira Rezende. 2014.
Using contextual information from topic hierarchies to
improve context-aware recommender systems.
Pattern Recognition (ICPR), 2014 22nd International
Conference on. IEEE, 3606–3611.</p>
      <p>Paul Grice. 1975. Logic and conversation. New York</p>
      <p>Academic Press, 41–58.</p>
      <p>Zhiheng Huang, Marcus Thint and Zengchang Qin.
2008. Question Classification Using Head Words
and Their Hypernyms. Proceedings of the
Conference on Empirical Methods in Natural Language
Processing. Association for Computational
Linguistics.</p>
      <p>James Lester, Karl Branting and Bradford Mott. 2004.</p>
      <p>Conversational agents. The Practical Handbook of
Internet Computing, 220–240.
of the eighth conference on European chapter of the
Association for Computational Linguistics.
Association for Computational Linguistics.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <string-name>
            <surname>John L. Austin</surname>
          </string-name>
          .
          <year>1962</year>
          .
          <article-title>How to Do Things With Words</article-title>
          . Oxford University Press.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          <string-name>
            <surname>James F. Allen</surname>
          </string-name>
          ,
          <string-name>
            <surname>Donna K. Byron</surname>
            , Myroslava Dzikovska, George Ferguson, Lucian Galescu and
            <given-names>Amanda</given-names>
          </string-name>
          <string-name>
            <surname>Stent</surname>
          </string-name>
          .
          <year>2001</year>
          .
          <article-title>Toward conversational human-computer interaction</article-title>
          .
          <source>AI magazine</source>
          ,
          <volume>22</volume>
          (
          <issue>4</issue>
          ):
          <fpage>27</fpage>
          -
          <lpage>39</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          <string-name>
            <given-names>Sabrina</given-names>
            <surname>Lombardi</surname>
          </string-name>
          ,
          <source>Sarabjot Singh Anand and Michele Gorgoglione</source>
          .
          <year>2009</year>
          .
          <article-title>Context and customer behaviour in recommendation.</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          <string-name>
            <given-names>Michael</given-names>
            <surname>Mauldin</surname>
          </string-name>
          .
          <year>1994</year>
          .
          <article-title>ChatterBots, TinyMuds, and the Turing Test: Entering the Loebner Prize Competition</article-title>
          .
          <source>AAAI Press - Proceedings of the Eleventh National Conference on Artificial Intelligence.</source>
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          <string-name>
            <given-names>Tomas</given-names>
            <surname>Mikolov</surname>
          </string-name>
          , Kai Chen, Greg Corrado and
          <string-name>
            <given-names>Jeffrey</given-names>
            <surname>Dean</surname>
          </string-name>
          .
          <year>2013</year>
          .
          <article-title>Efficient Estimation of Word Representations in Vector Space</article-title>
          .
          <source>arXiv:1301.3781 [cs.CL].</source>
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          <string-name>
            <given-names>Rambow</given-names>
            <surname>Owen</surname>
          </string-name>
          , Srinivas Bangalore and
          <string-name>
            <given-names>Marilyn</given-names>
            <surname>Walker</surname>
          </string-name>
          .
          <year>2001</year>
          .
          <article-title>Natural language generation in dialog systems</article-title>
          .
          <source>Proceedings of the first international conference on Human language technology research.</source>
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          <string-name>
            <given-names>Carston</given-names>
            <surname>Pask</surname>
          </string-name>
          .
          <year>1975</year>
          .
          <article-title>Conversation, cognition and learning</article-title>
          . New York: Elsevier.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          <string-name>
            <given-names>Diana</given-names>
            <surname>Perez-Marin</surname>
          </string-name>
          .
          <year>2011</year>
          .
          <article-title>Conversational Agents and Natural Language Interaction: Techniques and Effective Practices: Techniques and Effective Practices</article-title>
          . IGI Global.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          <string-name>
            <given-names>Silvia</given-names>
            <surname>Quarteroni</surname>
          </string-name>
          and
          <string-name>
            <given-names>Suresh</given-names>
            <surname>Manandhar</surname>
          </string-name>
          .
          <year>2007</year>
          .
          <article-title>A Chatbot-based Interactive Question Answering System</article-title>
          .
          <source>Proceedings of the 11th Workshop on the Semantics and Pragmatics of Dialogue</source>
          . Edited by Ron Artstein and
          <string-name>
            <given-names>Laure</given-names>
            <surname>Vieu</surname>
          </string-name>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          <string-name>
            <given-names>Verena</given-names>
            <surname>Rieser</surname>
          </string-name>
          and
          <string-name>
            <given-names>Oliver</given-names>
            <surname>Lemon</surname>
          </string-name>
          .
          <year>2011</year>
          .
          <article-title>Reinforcement learning for adaptive dialogue systems: a data-driven methodology for dialogue management and natural language generation</article-title>
          . Springer Science &amp; Business Media.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          <string-name>
            <given-names>Carston</given-names>
            <surname>Robyn</surname>
          </string-name>
          .
          <year>2002</year>
          .
          <article-title>Thoughts and Utterances: The Pragmatics of Explicit Communication</article-title>
          . Oxford Blackwell.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          <string-name>
            <given-names>Sophie</given-names>
            <surname>Rosset</surname>
          </string-name>
          , Delphine Tribout and
          <string-name>
            <given-names>Lori</given-names>
            <surname>Lamel</surname>
          </string-name>
          .
          <year>2008</year>
          .
          <article-title>Multi-level information and automatic dialog act detection in human-human spoken dialogs</article-title>
          .
          <source>Speech Communication</source>
          ,
          <volume>50</volume>
          (
          <issue>1</issue>
          ):
          <fpage>1</fpage>
          -
          <lpage>3</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          <string-name>
            <given-names>Francesco</given-names>
            <surname>Tarasconi</surname>
          </string-name>
          and Vittorio Di Tomaso.
          <year>2015</year>
          .
          <article-title>Geometric and statistical analysis of emotions and topics in corpora</article-title>
          .
          <source>IJCoL - Italian Journal of Computational Linguistics</source>
          , Vol.
          <volume>1</volume>
          n. 1. Accademia University Press.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          <string-name>
            <given-names>Chen</given-names>
            <surname>Xing</surname>
          </string-name>
          ,
          <string-name>
            <surname>Wei</surname>
            <given-names>Wu</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yu Wu</surname>
          </string-name>
          , Jie Liu, Yalou Huang,
          <string-name>
            <surname>Ming Zhou</surname>
          </string-name>
          , and
          <string-name>
            <surname>Wei-Ying Ma</surname>
          </string-name>
          .
          <year>2017</year>
          .
          <article-title>Topic Aware Neural Response Generation</article-title>
          . AAAI,
          <fpage>3351</fpage>
          -
          <lpage>3357</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          <string-name>
            <given-names>Gu</given-names>
            <surname>Yue</surname>
          </string-name>
          .
          <year>2017</year>
          .
          <article-title>Speech Intention Classification with Multimodal Deep Learning</article-title>
          .
          <source>30th Canadian Conference on Artificial Intelligence</source>
          ,
          <source>Canadian AI</source>
          <year>2017</year>
          , Proceeding. Springer.
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          <string-name>
            <given-names>Marilyn A.</given-names>
            <surname>Waler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Diane J.</given-names>
            <surname>Litman</surname>
          </string-name>
          , Candace A.
          <string-name>
            <surname>Kamm</surname>
            , and
            <given-names>Alicia</given-names>
          </string-name>
          <string-name>
            <surname>Abella</surname>
          </string-name>
          .
          <year>1997</year>
          .
          <article-title>PARADISE: A framework for evaluating spoken dialogue agents</article-title>
          .
          <source>Proceedings</source>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>