<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Adaptable Utterances in Voice User Interfaces to Increase Learnability</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Chelsea Myers</string-name>
          <email>chel.myers@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Anushay Furqan</string-name>
          <email>anushay.furqan@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Jichen Zhu</string-name>
          <email>jichen.zhu@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Author Keywords</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>ACM Classification Keywords</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Drexel University</institution>
          ,
          <addr-line>Philadelphia, PA 19104</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>H.5.2 [User Interfaces]: Voice I/O</institution>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Voice User Interface; Learnability; User Experience; Adaptive; Adaptable; Open User Models</institution>
        </aff>
      </contrib-group>
      <fpage>44</fpage>
      <lpage>49</lpage>
      <abstract>
        <p>Voice User Interfaces (VUIs) are growing in popularity as a method of controlling smart home features. However, as VUIs grow in popularity, major obstacles still negatively impact their performance and user experience. Since VUIs are invisible by nature, users find it difficult to learn the supported features and verbal commands. Discovering the correct verbal commands is made more difficult since users have a verbiage preference and developers of VUIs cannot pre-program all possible commands. We propose adaptable verbal commands, termed adaptable utterances, and Open User Models (OUMs) as a method to allow customization of a VUI's commands to match the individual user's preference. We review relevant research on adaptive and adaptable VUIs and identify the limitations adaptable utterances and OUMs could address. Finally, we present a sample study design to evaluate our proposed methods.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>Copyright © 2018 for this paper held by its author(s). Copying permitted for private
and academic purposes.</p>
    </sec>
    <sec id="sec-2">
      <title>Introduction</title>
      <p>
        Voice User Interfaces (VUIs) started becoming more
common in everyday objects with the release of iPhone’s Siri
and the Google Assistant on Google phones. In addition
to the VUIs embedded in smart phones, standalone smart
objects, such as Amazon’s Echo and Google Home, made
VUIs more popular because of their accessibility and error
prevention [
        <xref ref-type="bibr" rid="ref23">23</xref>
        ]. As smart objects and smart homes become
increasingly popular, we believe VUIs will be an important
interaction method for them. However, the invisible nature
of VUIs and their high cognitive load are major obstacles.
Users struggle to learn VUI’s intents (i.e. features) and the
utterances (i.e. verbal commands) that trigger them [
        <xref ref-type="bibr" rid="ref20 ref22">22,
20</xref>
        ]. This issue can compound if standalone VUIs become a
hub on control for smart homes.
      </p>
      <p>
        In this paper, we propose to explore the use of adaptation
and open user models (OUM) to make VUIs more
transparent and learnable. OUMs are user models accessible
to the user of a system. Ahn [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] defines OUMs as aiming
to “transparently reveal the personalization process so that
users can easily understand the internal mechanisms and
anticipate the behavior of the system." In existing
literature, techniques for increasing a VUI’s learnability through
starter tutorials and in-app tutorials have been explored
[
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. Adaptive and adaptable techniques are often applied
to VUIs to support usability and learnability. Although VUI
research focuses more on adaptive techniques [
        <xref ref-type="bibr" rid="ref10 ref14 ref16 ref17 ref6 ref7">14, 6, 16,
17, 10, 7</xref>
        ], adaptive VUIs have the potential to confuse their
users if the correct mental model of the adaptation is not
formed [
        <xref ref-type="bibr" rid="ref10 ref15">15, 10</xref>
        ].
      </p>
      <p>
        We propose an adaptable VUI with a GUI OUM that
allows user to edit the utterances a VUI accepts.
Developers of VUIs cannot pre-program all the possible utterances
to meet the preferences of each individual users. We
believe that adaptable utterances will allow users to customize
the verbiage of their VUI to match their preferences. We
hypothesize that the GUI OUM of the VUI’s intents and
utterances, coupled with the act of adapting utterances, will
increase the transparency of the VUI and its learnability.
In this paper, we present our adaptable utterance and OUM
techniques in relation to adaptive and adaptable VUI
research. We identify the limitations that adaptable utterances
could potentially address. To evaluate the impact of
adaptable utterances, we propose a study design for three
different adaptable techniques: adaptable tutorials, OUM
adaptation, and in-app adaptation. We have designed a
calendarmanagement VUI called DiscoverCal (Figures 1 &amp; 2) that
has been evaluated in previous studies[
        <xref ref-type="bibr" rid="ref10 ref20">10, 20</xref>
        ]. DiscoverCal
runs on a wall-mounted display, with voice control being its
primary input. Our study design incorporates these
adaptable techniques into DiscoverCal and a companion website.
The rest of the paper is structured as follows. First, we
review related adaptable VUI research and then we present
our methodology for this paper’s study.
      </p>
    </sec>
    <sec id="sec-3">
      <title>Adaptive and Adaptable VUIs</title>
      <p>
        Research explores different approaches to personalize
VUIs for increased learnability and reducing the user’s
cognitive load. Among these approaches are creating an
adaptive or an adaptable VUI system. Adaptive VUI systems
are personalized automatically to the user by adapting the
system’s feedback [
        <xref ref-type="bibr" rid="ref15 ref19">15, 19</xref>
        ], initiative [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ], context [
        <xref ref-type="bibr" rid="ref12 ref21">21, 12</xref>
        ],
and/or visual aids [
        <xref ref-type="bibr" rid="ref10 ref7">10, 7</xref>
        ]. Adaptable VUI systems are
personalized, but only through the user’s initiative. Users can
change the settings of a VUI to their preference.
      </p>
      <sec id="sec-3-1">
        <title>Adaptive VUIs</title>
        <p>
          Popular adaptive VUI techniques alter feedback and
initiative. Karsenty and Botherel [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ] developed a VUI,
TRAVELS, with a guided and unguided mode that relies on
changing the initiative of the system and the amount of feedback
provided. Adaptation occurs after users were deemed no
longer novices with the system. Karsenty and Botherel
found that the switch to the unguided mode increased
system errors temporarily. They observed that “several users
were confused by the system’s change of behavior." [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ]
Research has found adaptive techniques can increase
usability of VUI research [
          <xref ref-type="bibr" rid="ref16 ref17 ref6">16, 17, 6</xref>
          ], but if the correct mental
model of the adaptation is not formed by the user, it can be
confusing [
          <xref ref-type="bibr" rid="ref10 ref15">15, 10</xref>
          ]. In a previous study with DiscoverCal we
explored using a visual adaptive menu to show more
complex commands for the system as the user’s expertise grew
(Figure 2). We found that several users using the adaptive
version of DiscoverCal were confused by the adaptation
and were not sure why the menu changed. This
observation lead us to explore other tools to support
learnability and transparency for VUIs. Adaptive techniques are a
useful tool for VUI design. However, we believe adaptable
techniques are under-explored and can address the
transparency issues that adaptive techniques pose as seen with
        </p>
      </sec>
      <sec id="sec-3-2">
        <title>TRAVELS and DiscoverCal.</title>
      </sec>
      <sec id="sec-3-3">
        <title>Adpatable VUIs</title>
        <p>
          Dusan and Flanagan [
          <xref ref-type="bibr" rid="ref8 ref9">8, 9</xref>
          ] evaluated adaptable intents and
utterances with a multimodal VUI drawing application,
using a digital touch display with pen and voice input.
Participants could create a new intents and utterances for colors,
shapes, and other features of the drawing app. Although
this study does include adaptable utterances, the
application required intensive training to use and memorization of
a programmatic language to create intents and utterances.
The evaluation also focused on the system’s technical
performance and not its learnability. To our knowledge, there is
no empirical evaluation on how adaptable VUI techniques
can be used to allow the user to alter a VUI system’s
utterances to increase learnability. Adaptable research instead
focuses on initiative [
          <xref ref-type="bibr" rid="ref18">18</xref>
          ], content words, and self-learning
VUIs.
        </p>
      </sec>
      <sec id="sec-3-4">
        <title>Adaptable Initiative</title>
        <p>
          VUI initiative is determined by who is leading the
conversation. If the VUI is asking the user questions, the VUI has the
initiative (and vice-versa). Adaptable VUI research so far
has empirically evaluated an adaptable and non-adaptable
initiative of a VUI and found the adaptable version
"outpreformed" the non-adaptable [
          <xref ref-type="bibr" rid="ref18">18</xref>
          ]. Users of the
adaptive version were allowed to swap the initiative of the VUI
themselves to navigate through errors. Although this VUI is
adaptable, the user only has control over the initiative. We
propose to evaluate the affect of adaptable utterances on a
VUI’s learnability.
        </p>
      </sec>
      <sec id="sec-3-5">
        <title>Adaptable Content Words</title>
        <p>
          Content words are the entities in an utterance. They are
not the keywords that map to an intent, but instead provide
information. In the utterance, “Add an event titled Lunch
located at the cafe," “Lunch" and “the cafe" are content
words. The rest are the keywords that would be detected
by a VUI and mapped to their corresponding intent.
Adaptable research explores increasing the vocabulary of VUIs
by adding unidentified content words. Seneff et al. [
          <xref ref-type="bibr" rid="ref21">21</xref>
          ]
developed a VUI that automatically detected out-of-vocabulary
words (OOV). The user could then choose to add that OOV
word to the system by spelling it out. Brill [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ] developed an
algorithm for NLP that also detected OOV words but would
guess their category. The user could confirm this
categorization (e.g. labeling “the cafe" as a restaurant). Both of
these studies do not evaluate their system with users and
instead test with text corpora. We aim to take this adaptable
technique further by allowing users to also change the
keywords in an utterance that maps to an intent and evaluate
its impact on the learnability.
        </p>
      </sec>
      <sec id="sec-3-6">
        <title>Self-Learning VUIs</title>
        <p>
          Relevant research allowing users to generate their own
utterances includes self-learning VUIs like ALADIN [
          <xref ref-type="bibr" rid="ref11 ref13">11, 13</xref>
          ].
ALADIN is used to control smart homes for disabled users
and/or users with dysarthria speech. ALADIN has no
vocabulary, grammar, or feature libraries initially, and instead
is completely trained by the user on first use and
continually through further usage. However, this type of research
has only been evaluated in a demonstration setting. No
empirical analysis has been completed on the implications of
a self-learning VUI on it’s users’ experience. Our research
differs by examining a VUI that does have pre-defined
vocabulary, grammar, and feature libraries; mirroring modern
VUIs. We aim to analyze how learnability is impacted by
allowing users to change the VUI’s utterances, while keeping
the intents the same.
        </p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Open User Models</title>
      <p>
        OUMs have been found to be generally beneficial for
adaptive systems [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. Open Student Models, a category of OUMs,
are used in adaptive personalized E-Learning to encourage
students to reflect on their progress, strengths, and
weaknesses [
        <xref ref-type="bibr" rid="ref4 ref5">5, 4</xref>
        ]. OUMs provide a layer of transparency for the
user to view the information stored about them. Because
of the benefits OUMs have been observed to provide, we
propose to evaluate the impact of an OUM on a VUI’s
learnability. Since users of our OUM will be able to edit
utterances, our OUM can be categorized as an editable OUM.
An editable OUM’s adaptive search/filtering system, similar
to our adaptable utterances, was initially found to hurt the
system’s performance [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. The study was later redone and it
was found that the design of the OUM was too complicated
and user edits to the OUM negatively affected performance.
The editable OUM was redesigned to be more visual, and
researchers found the OUM positively impacted
performance [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. From this, we learn it is important to design a
clear OUM where the user is aware of the information
presented and how to manipulate it.
      </p>
    </sec>
    <sec id="sec-5">
      <title>Study Design</title>
      <p>To analyze adaptable utterances and OUMs, we propose a
between-subject study evaluating three adaptable utterance
techniques with DiscoverCal compared to non-adaptive
version focusing on extended learnability. The three adaptable
techniques are as follows: adaptable tutorials, OUM
adaptation, and in-app adaptation.</p>
      <p>Adaptable Tutorials - Tutorials are a common learning tool
for VUIs. Our technique creates a tutorial for DiscoverCal’s
VUI that also includes utterance“training". In this training,
users will declare what utterances they wish to associate to
each intent. This is similar to ALADIN’s self-learning
training, but users can only change utterances, not the intents</p>
      <sec id="sec-5-1">
        <title>DiscoverCal supports.</title>
        <p>OUM Adaptation - Allowing users to view the OUM in
DiscoverCal’s companion website. Here users can update,
add, and delete utterances for each intent.</p>
        <p>In-app Adaptation - Allowing users to update utterances
while using DiscoverCal (not in the OUM) to their
preference.</p>
        <p>We have selected these three techniques to evaluate the
impact of different adaptable utterance methods that
occur in different parts of a VUI: the tutorial, OUM, and in the
application. Recruited participants would be balanced for
technical skill and previous VUI experience. Each
participant would interact with DiscoverCal and perform a select
set of tasks. Tasks would have the participants add, edit,
and delete events from DiscoverCal. To analyze the impact
on extended learnability, the study would be broken into
three sessions over the course of one week. Performance
metrics (e.g. time spent per task, total time per session,
errors encounter) will be collected and a subjective usability
completed per session to measure extended learnability.</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>Conclusion</title>
      <p>In this paper we propose the evaluation of adaptable
utterances and OUMs as a tool for increasing the learnability
and transparency of VUIs. We next present an overview
of adaptive and adaptable techniques and their limitations.
Although adaptive VUI techniques generally benefit a VUI,
the dynamic adaptation can confuse users. We hypothesize
that adaptable utterances and OUMs can customize a VUI’s
vocabulary to the preference of the user while avoiding this
confusion. Finally, we propose a sample study design to
evaluate these methods.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Jae-wook Ahn</surname>
          </string-name>
          , Peter Brusilovsky, Jonathan Grady, Daqing He, and Sue Yeon Syn.
          <year>2007</year>
          .
          <article-title>Open User Profiles for Adaptive News Systems: Help or Harm?</article-title>
          .
          <source>In Proceedings of the 16th International Conference on World Wide Web (WWW '07)</source>
          . ACM, New York, NY, USA,
          <fpage>11</fpage>
          -
          <lpage>20</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Jae-wook Ahn</surname>
          </string-name>
          ,
          <string-name>
            <surname>Peter Brusilovsky</surname>
          </string-name>
          , and Shuguang Han.
          <year>2015</year>
          .
          <article-title>Personalized search: reconsidering the value of open user models</article-title>
          .
          <source>In Proceedings of the 20th International Conference on Intelligent User Interfaces. ACM</source>
          ,
          <volume>202</volume>
          -
          <fpage>212</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <given-names>Eric</given-names>
            <surname>Brill</surname>
          </string-name>
          .
          <year>1995</year>
          .
          <article-title>Transformation-based error-driven learning and natural language processing: A case study in part-of-speech tagging</article-title>
          .
          <source>Computational linguistics 21</source>
          ,
          <issue>4</issue>
          (
          <year>1995</year>
          ),
          <fpage>543</fpage>
          -
          <lpage>565</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <given-names>Susan</given-names>
            <surname>Bull</surname>
          </string-name>
          .
          <year>2004</year>
          .
          <article-title>Supporting learning with open learner models</article-title>
          .
          <source>Planning</source>
          <volume>29</volume>
          ,
          <issue>14</issue>
          (
          <year>2004</year>
          ),
          <fpage>1</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <given-names>Susan</given-names>
            <surname>Bull</surname>
          </string-name>
          , Paul Brna, and
          <string-name>
            <given-names>Helen</given-names>
            <surname>Pain</surname>
          </string-name>
          .
          <year>1995</year>
          .
          <article-title>Extending the scope of the student model</article-title>
          .
          <source>User Modeling and User-Adapted Interaction 5</source>
          ,
          <issue>1</issue>
          (
          <year>1995</year>
          ),
          <fpage>45</fpage>
          -
          <lpage>65</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <given-names>J.</given-names>
            <surname>Chu-Carroll</surname>
          </string-name>
          .
          <year>2000</year>
          .
          <article-title>MIMIC: An adaptive mixed initiative spoken dialogue system for information queries</article-title>
          .
          <source>Proceedings of the sixth conference on Applied natural language processing 6</source>
          (
          <year>2000</year>
          ),
          <fpage>97</fpage>
          -
          <lpage>104</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <given-names>Eric</given-names>
            <surname>Corbett</surname>
          </string-name>
          and
          <string-name>
            <given-names>Astrid</given-names>
            <surname>Weber</surname>
          </string-name>
          .
          <year>2016</year>
          .
          <article-title>What can I say? Proceedings of the 18th International Conference on Human-Computer Interaction with Mobile Devices</article-title>
          and Services - MobileHCI '
          <volume>16</volume>
          (
          <year>2016</year>
          ),
          <fpage>72</fpage>
          -
          <lpage>82</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <given-names>S.</given-names>
            <surname>Dusan</surname>
          </string-name>
          and
          <string-name>
            <given-names>J.</given-names>
            <surname>Flanagan</surname>
          </string-name>
          .
          <year>2002</year>
          .
          <article-title>Adaptive dialog based upon multimodal language acquisition</article-title>
          .
          <source>Proceedings - 4th IEEE International Conference on Multimodal Interfaces</source>
          ,
          <string-name>
            <surname>ICMI</surname>
          </string-name>
          <year>2002</year>
          (
          <year>2002</year>
          ),
          <fpage>135</fpage>
          -
          <lpage>140</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <given-names>Sorin</given-names>
            <surname>Dusan and James L Flanagan</surname>
          </string-name>
          .
          <year>2001</year>
          .
          <article-title>Human language acquisition by computers</article-title>
          .
          <source>In Proceedings of the International Conference on Robotics, Distance Learning and Intelligent Communication Systems</source>
          .
          <volume>387</volume>
          -
          <fpage>392</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Anushay</surname>
            <given-names>Furqan</given-names>
          </string-name>
          , Chelsea Myers, and
          <string-name>
            <given-names>Jichen</given-names>
            <surname>Zhu</surname>
          </string-name>
          .
          <year>2017</year>
          .
          <article-title>Learnability through Adaptive Discovery Tools in Voice User Interfaces</article-title>
          .
          <source>Proceedings of the 2017 CHI Conference Extended Abstracts on Human Factors in Computing Systems - CHI EA '17</source>
          (
          <year>2017</year>
          ),
          <fpage>1617</fpage>
          -
          <lpage>1623</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Jort F. Gemmeke</surname>
          </string-name>
          , Bart Ons, Netsanet Tessema, Hugo Van Hamme,
          <string-name>
            <surname>Janneke Van De Loo</surname>
          </string-name>
          , Guy De Pauw, Walter Daelemans, Jonathan Huyghe, Jan Derboven, Lode Vuegen, Bert Van Den Broeck,
          <string-name>
            <surname>Peter Karsmakers</surname>
            , and
            <given-names>Bart</given-names>
          </string-name>
          <string-name>
            <surname>Vanrumste</surname>
          </string-name>
          .
          <year>2013</year>
          .
          <article-title>Self-taught assistive vocal interfaces: An overview of the ALADIN project</article-title>
          .
          <source>Proceedings of the Annual Conference of the International Speech Communication Association</source>
          , INTERSPEECH (
          <year>2013</year>
          ),
          <fpage>2039</fpage>
          -
          <lpage>2043</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <given-names>James</given-names>
            <surname>Glass</surname>
          </string-name>
          and
          <string-name>
            <given-names>Stephanie</given-names>
            <surname>Seneff</surname>
          </string-name>
          .
          <year>2003</year>
          .
          <article-title>Flexible and personalizable mixed-initiative dialogue systems</article-title>
          .
          <source>Proceedings of the HLT-NAACL 2003 workshop on Research directions in dialogue processing - 7</source>
          (
          <year>2003</year>
          ),
          <fpage>19</fpage>
          -
          <lpage>21</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Jonathan</surname>
            <given-names>Huyghe</given-names>
          </string-name>
          , Jan Derboven, and Dirk De Grooff.
          <year>2014</year>
          .
          <article-title>ALADIN: Demo of a Multimodal Adaptive Voice Interface</article-title>
          .
          <source>Proceedings of the 8th Nordic Conference on Human-Computer Interaction: Fun</source>
          , Fast, Foundational (
          <year>2014</year>
          ),
          <fpage>1035</fpage>
          -
          <lpage>1038</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <given-names>Pontus</given-names>
            <surname>Johansson</surname>
          </string-name>
          .
          <year>2002</year>
          .
          <article-title>NLP Techniques for Adaptive Dialogue Systems</article-title>
          . Techniques (
          <year>2002</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <given-names>Laurent</given-names>
            <surname>Karsenty</surname>
          </string-name>
          and
          <string-name>
            <given-names>Valerie</given-names>
            <surname>Botherel</surname>
          </string-name>
          .
          <year>2005</year>
          .
          <article-title>Transparency strategies to help users handle system errors</article-title>
          .
          <source>Speech Communication</source>
          <volume>45</volume>
          ,
          <article-title>3 SPEC</article-title>
          . ISS. (
          <year>2005</year>
          ),
          <fpage>305</fpage>
          -
          <lpage>324</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <surname>Diane</surname>
            J Litman and
            <given-names>Shimei</given-names>
          </string-name>
          <string-name>
            <surname>Pan</surname>
          </string-name>
          .
          <year>2000</year>
          .
          <article-title>Predicting and adapting to poor speech recognition in a spoken dialogue system</article-title>
          .
          <source>Proceedings of the National Conference on Artificial Intelligence</source>
          (
          <year>2000</year>
          ),
          <fpage>722</fpage>
          -
          <lpage>728</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          17.
          <string-name>
            <surname>Diane</surname>
            J Litman and
            <given-names>Shimei</given-names>
          </string-name>
          <string-name>
            <surname>Pan</surname>
          </string-name>
          . 2002a.
          <article-title>Designing and Evaluating an Adaptive Spoken Dialogue System</article-title>
          . (
          <year>2002</year>
          ),
          <fpage>111</fpage>
          -
          <lpage>137</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          18.
          <string-name>
            <surname>Diane</surname>
            J Litman and
            <given-names>Shimei</given-names>
          </string-name>
          <string-name>
            <surname>Pan</surname>
          </string-name>
          . 2002b.
          <article-title>Designing and Evaluating an Adaptive Spoken Dialogue System</article-title>
          . (
          <year>2002</year>
          ),
          <fpage>111</fpage>
          -
          <lpage>137</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          19.
          <string-name>
            <given-names>François</given-names>
            <surname>Mairesse</surname>
          </string-name>
          and
          <string-name>
            <given-names>Marilyn</given-names>
            <surname>Walker</surname>
          </string-name>
          .
          <year>2005</year>
          .
          <article-title>Learning to Personalize Spoken Generation for Dialogue Systems</article-title>
          .
          <source>Proceedings of Interspeech'2005 - Eurospeech: 9th European Conference on Speech Communication and Technology October</source>
          (
          <year>2005</year>
          ),
          <fpage>1881</fpage>
          -
          <lpage>1884</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          20.
          <string-name>
            <surname>Chelsea</surname>
            <given-names>Myers</given-names>
          </string-name>
          , Jessica Nebolsky, Karina Caro, and
          <string-name>
            <given-names>Jichen</given-names>
            <surname>Zhu</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>Patterns for How Users Overcome Obstacles in Voice User Interfaces</article-title>
          .
          <article-title>(</article-title>
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          21.
          <string-name>
            <surname>Stephanie</surname>
            <given-names>Seneff</given-names>
          </string-name>
          , Grace Chung, and
          <string-name>
            <given-names>Chao</given-names>
            <surname>Wang</surname>
          </string-name>
          .
          <year>2003</year>
          .
          <article-title>Empowering End Users to Personalize Dialogue Systems through Spoken Interaction 1</article-title>
          .
          <string-name>
            <given-names>Processing</given-names>
            <surname>January</surname>
          </string-name>
          (
          <year>2003</year>
          ),
          <fpage>749</fpage>
          -
          <lpage>752</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          22.
          <string-name>
            <given-names>Ben</given-names>
            <surname>Shneiderman</surname>
          </string-name>
          .
          <year>2000</year>
          .
          <article-title>The limits of speech recognition</article-title>
          .
          <source>Commun. ACM 43</source>
          ,
          <issue>9</issue>
          (
          <year>2000</year>
          ),
          <fpage>63</fpage>
          -
          <lpage>65</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          23.
          <string-name>
            <given-names>Kathryn</given-names>
            <surname>Whitenton</surname>
          </string-name>
          .
          <year>2016</year>
          .
          <string-name>
            <given-names>Voice</given-names>
            <surname>Interaction</surname>
          </string-name>
          <string-name>
            <surname>UX</surname>
          </string-name>
          : Brave New World...Same Old Story. (
          <year>2016</year>
          ). https://www. nngroup.com/articles/voice-interaction-ux/
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>