=Paper= {{Paper |id=Vol-2903/IUI21WS-HAIGEN-12 |storemode=property |title=Creative PenPal: A Virtual Embodied Conversational AI Agent to Improve User Engagement and Collaborative Experience in Human-AI Co-Creative Design Ideation |pdfUrl=https://ceur-ws.org/Vol-2903/IUI21WS-HAIGEN-12.pdf |volume=Vol-2903 |authors=Jeba Rezwana,Mary Lou Maher,Nicholas Davis |dblpUrl=https://dblp.org/rec/conf/iui/RezwanaMD21 }} ==Creative PenPal: A Virtual Embodied Conversational AI Agent to Improve User Engagement and Collaborative Experience in Human-AI Co-Creative Design Ideation== https://ceur-ws.org/Vol-2903/IUI21WS-HAIGEN-12.pdf
Creative PenPal: A Virtual Embodied Conversational AI
Agent to Improve User Engagement and Collaborative
Experience in Human-AI Co-Creative Design Ideation
Jeba Rezwanaa , Mary Lou Mahera and Nicholas Davisa
a
    University of North Carolina at Charlotte, 9201 University City Blvd, Charlotte, NC 28223, US


                                             Abstract
                                             In recent years, researchers have designed many co-creative systems that are very promising with a powerful AI, yet some
                                             fail to engage the users due to the unimpressive quality of the collaboration and interaction. Most of the existing co-creative
                                             systems use instructing interaction where users only communicate with the AI by providing instructions for contribution.
                                             In this paper, we demonstrate the prototype of a co-creative system for design ideation, Creative PenPal that utilizes an
                                             interaction model that includes human-AI conversing interaction using text and a virtual embodiment of the AI character. We
                                             hypothesize that this interaction model will improve user engagement, user perception about the AI, and the collaborative
                                             experience. We describe the study design to investigate the impact of this particular interaction model on user engagement
                                             and the overall collaborative experience. By the time of the workshop, we will have the data and insights from the study.

                                             Keywords
                                             Human AI Co-creation, User Engagement, Virtual Embodied AI, Conversational AI



1. Introduction                                                                                                          With the intention of investigating the trends in the
                                                                                                                      interaction design of the existing co-creative systems, we
AI agents are becoming a part of our everyday life, thanks                                                            utilized the archival website, ”Library of Mixed-Initiative
to artificial intelligence technologies. Human-AI co cre-                                                             Creative Interfaces” (LMICI), which archives 74 exist-
ativity involves a human and an AI collaborating on cre-                                                              ing co-creative systems [5]. Angie Spoto and Natalia
ative tasks as partners [1]. Rather than being perceived as                                                           Oleynik created this archive after a workshop on mixed-
a support tool, AI agents in co-creative systems should be                                                            initiative interfaces led by Deterding et al. in 2017 [6, 5].
regarded as a co-equal partner. This field has the poten-                                                             The archive provides the corresponding literature and
tial to transform how people perceive and interact with                                                               other relevant information for each of the systems. We
AI. A study showed that AI ability alone does not ensure                                                              analyzed the interaction designs of the co-creative sys-
a positive collaborative experience of users with the AI                                                              tems present in the archive. Apparently, most of the co-
[2]. In recent years, researchers have designed many                                                                  creative systems use instructing interaction [7], where
co-creative systems with powerful AI ability, yet some-                                                               users provide instructions to the system using buttons,
times users fail to maintain their interest and engagement                                                            sliders, or text to communicate directly with the AI (other
while collaborating with the AI due to the quality of the                                                             than communicating through the creative product). How-
collaboration and interaction. The literature asserts that                                                            ever, using buttons and sliders, users can communicate
user engagement is associated with the way users interact                                                             with the AI in a very constrained and minimal way in
with a system [3]. Interaction design is often an untended                                                            most of the systems. Very few systems use text, voice,
topic in the co-creativity literature despite being a funda-                                                          or embodied communication for user to AI direct com-
mental property of co-creative systems. Bown asserted                                                                 munication during a co-creation to provide information
that the success of a creative system’s collaborative role                                                            to the AI, give feedback to the AI, etc. For example, Im-
should be further investigated through interaction design                                                             age to Image [8] is a co-creative system that converts a
as interaction plays a key role in the creative process of                                                            line drawing of a particular object from the user into a
co-creative systems [4]. Therefore, as a young field, there                                                           photo-realistic image. The user interface has only one
are potential areas of interaction design to be explored                                                              button that users click to tell the AI to convert the draw-
for designing effective co-creative systems that engage                                                               ing. Other than the button, there is no way of commu-
users and provide a better collaborative experience.                                                                  nicating with the AI to provide information, suggestion
                                                                                                                      or feedback. In a human collaboration, collaborators
Joint Proceedings of the ACM IUI 2021 Workshops, College Station,                                                     communicate to provide feedback and convey important
USA                                                                                                                   information to each other and is a major component of
Envelope-Open jrezwana@uncc.edu (J. Rezwana); m.maher@uncc.edu                                                        the mechanics of co-creation [9]. The literature about
(M. L. Maher); ndavis64@uncc.edu (N. Davis)                                                                           human-AI co-creation says that embodied communica-
                                       © 2021 Copyright for this paper by its authors. Use permitted under Creative
    CEUR
    Workshop
                  http://ceur-ws.org
                  ISSN 1613-0073
                                       Commons License Attribution 4.0 International (CC BY 4.0).
                                       CEUR Workshop Proceedings (CEUR-WS.org)
                                                                                                                      tion improves coordination between the human and the
    Proceedings
AI [10]. Additionally, literature asserts that a commu-       buttons or using function keys, etc. In contrast, the con-
nication channel for conversation between co-creators         versing interaction type is where users have a dialogue
other than communicating through the shared creative          with a system. Users can speak via an interface or type in
product improves user engagement in a human creative          questions or answers to which the system replies via text
collaboration [11]. These literatures led us to investi-      or speech output [13]. Conversational agents have tran-
gate the impact of embodied communication from the AI         sitioned into multiple industries with increased ability
and a conversation between the human and AI on user           for user engagement in intelligent conversation.
engagement and collaborative experience in human-AI              The literature asserts that embodied communication
co-creativity. Our research questions emerged from the        aids synchronization and coordination in improvisational
issue that most existing co-creative systems use instruct-    human-computer co-creativity [10]. Being able to con-
ing interaction type, which uses one-way communica-           verse with each other shows an increased engagement
tion, human to AI. For this work, we will investigate         level in a human creative collaboration [11]. A user’s
the impact of conversing interaction and AI embodiment        confidence in an AI agent’s ability to perform tasks is
on user engagement, user perception about the AI and          improved when imbuing the agent with embodiment and
the overall collaborative experience. The two research        social behaviors compared to the agent solely depending
questions we have are-                                        on conversation [14]. Bente et al. reported that em-
                                                              bodied telepresent communication improved both social
     • How does AI embodiment and conversing inter-           presence and interpersonal trust in remote collaboration
       action influence user engagement?                      settings with a high level of nonverbal activity [15].
     • How does AI embodiment and conversing interac-            User engagement with virtual embodied conversational
       tion influence user perception about the AI agent      agents can be measured via user self-reports; by mon-
       as the collaborative partner and the overall col-      itoring the user’s responses, tracking the user’s body
       laborative experience?                                 postures, head movements and facial expressions dur-
                                                              ing the interaction, or by manually logging behavioral
   For investigating the research questions, we have de-      responses of user experience [16]. Carrol and Latulipe
veloped a prototype of a co-creative system named Cre-        proposed a quantitative and psychometric survey, called
ative PenPal where the user and the AI collaborate on a       Creativity Support Index (CSI), to assess a tool’s creativ-
design ideation task. Users can generate ideas for design-    ity support by measuring six dimensions of creativity
ing a particular object by sketching on a canvas, and the     via self reports: Exploration, Expressiveness, Immersion,
AI will also contribute to the design ideation by showing     Enjoyment, Results Worth Effort and Collaboration [17].
different inspirational sketches. Creative PenPal utilizes
a conversing interaction for the communication between
the human and the AI. Additionally, a virtual embodied
character for the AI agent is utilized. For investigating 3. Interface
the research questions, we describe the study design in
                                                              Creative PenPal is an interactive prototype, created with
the paper. By the time of the workshop, we will have the
                                                              Javascript, which has all the interaction components ex-
data and insights from the study.
                                                              cept the back-end AI model. We have selected a collection
                                                              of sketches as the database for creating a seamless expe-
2. Related Work                                               rience that mimics an actual implementation of the AI
                                                              model. The sketch generation is automated where the
Louie et al. identified that AI ability alone does not ensure
                                                              system selects sketches from the collection. We have two
a positive collaborative experience of users with the AI
                                                              versions of the Creative PenPal prototype to investigate
[2]. Bown asserted that the success of a creative system’s
                                                              and compare the user engagement and collaborative ex-
collaborative role should be further investigated in terms
                                                              perience between the two versions. The original version
of interaction design as interaction plays a key role in
                                                              uses a conversing interaction and a virtual embodied AI
the creative process of co-creative systems [4]. Later Yee-
                                                              (see Figure 1). The virtual embodied AI character, a pen-
King and d’Inverno argued for a stronger focus on the
                                                              cil, is shown in section A of Figure 1. We will address
user experience, suggesting a need for further integration
                                                              the AI character as PenPal in the rest of the paper. Sec-
of interaction design practice into human-AI co-creativity
                                                              tion B is where the conversation happens between the
research [12].
                                                              PenPal and the user via text and buttons. We can see
   Interaction types are ways a user interacts with a prod-
                                                              the design task displayed in section C. Both the user and
uct or application [13]. Instructing interaction is where
                                                              the AI collaborate in a design ideation task where both
users issue instructions to a system. This can be done
                                                              collaborators generate ideas for the design of an object as
in many ways, including typing in commands, selecting
                                                              sketches. Users will design the specified object in the task
options from menus or on a multitouch screen, pressing
                                                 A                                                             B
                              C

                                  D                                             E




                                                      F                                                            G




Figure 1: Original Creative PenPal Interface with AI Embodiment and Conversing Interaction


by sketching on the canvas shown in section F. Users can      its contribution, it will be seen as happy and when the
undo a stroke using the ”Undo Previous Sketch” button         user does not like the contribution from the AI, it will
and start the design ideation over by using the ”Clear the    be sad. The conversation is divided into five different
canvas” button. When users hit the ”Inspire me” button        situational phases demonstrated in Figure 2. Each phase
shown in section B, the virtual AI character will show        includes the embodied state of the AI and conversational
an inspirational sketch of a conceptually similar object,     interaction between the user and the PenPal. The text
an object that have similar working mechanism or usage        without a comment bubble represents the embodied state
as the design task object, on its canvas shown in sec-        of the AI in Figure 2. The texts with comment bubbles
tion G. Previous work on co-creative design ideation [18]     represent dialogues of the user and the AI, and the icon
showed that users were more inspired by conceptually          indicates which dialogue belongs to whom. Different
similar objects than visually similar objects that share      responses from the user initiate another phase, which is
structural similarity as the design task object. Users can    shown using arrows in Figure 2. If the user can respond
also ask for visually similar objects or sketches of the      with different options, “/” sign is used in the Figure.
design task object to get inspiration by saying they didn’t
like the conceptually similar object (described in the next   4.1. PenPal Introduction
section). Section E shows the name of the object located      This phase will start when the user starts the design
in the PenPal generated sketch. The other version uses        task. PenPal will introduce itself and ask the users if
an instructing interaction where users can instruct the       they want to see an inspirational sketch from the AI by
AI using buttons without AI embodiment (Figure 2). We         saying ”Hi! I am your Creative PenPal. Do you want
will use both of these two versions to compare the impact     me to inspire you?”. Users can respond immediately by
of two different interaction designs on user engagement       pressing the button ”Inspire me” or they can keep ideating
and collaborative experience with an AI.                      by sketching and respond later.

4. Interaction Model                                          4.2. PenPal Generating Sketch and
                                                                   Collecting User Preferences
For the interaction model, we choose a conversing in-
teraction. The conversation with the virtual embodied         When the user hits the button ”Inspire me” indicating the
AI is simple so that the user will be able to go deeper       desire to see an inspirational sketch, the PenPal will move
into the ideation process without any interruption in the     to the canvas and generate a sketch. The PenPal will ask
design flow. The embodied virtual agent will show some        the user whether they liked the sketch or not. The user
affective characteristics, for example, when the user likes
Figure 2: Creative PenPal Interface with Instructing Interaction and without AI Embodiment




Figure 3: Conversation Model between PanPal and the User


can reply with the ”Yes” button or the ”No” button. This    4.3. User Liked PenPals Sketch
phase is for collecting user preferences.
                                                            When users select the ”Yes” button in response to Pen-
                                                            Pal’s question to determine whether the user liked the
                                                            sketch or not, it means the sketch inspired the user in
Figure 4: PenPal Collecting User Preference


their design ideation. The PenPal will arrive with a happy      including 25 males and 25 females as participants. This
face and say, ”I am glad that you liked the sketch! Let me      study will use a between-subject study where one group
know if you want to see another inspirational sketch as         of participants will test the version with instructing in-
an idea”. If The user wants to see an inspiration again,        teraction and without any embodied AI character. The
they will select the ”Inspire me” button.                       other group will test the version with the conversing
                                                                interaction and a virtual embodied AI agent. The study
4.4. User Did not Like PenPals Sketch                           will start with a short pre-study survey to collect some
When users click the ”No” button, indicating that Pen-          demographic information about the participants, for ex-
Pal’s generated sketch did not inspire them, PenPal ar-         ample, gender, age-range, drawing/ sketching skills, etc.
rives with a sad face and says, ”Sorry that I could not         Then, the participant will carry out the design task using
inspire you!” (left side of Figure 4, the gree arrow indi-      either one version of the Creative PenPal. The task for
cates transition). Then it suggests the user ask for specific   this study is- “Ideate the design of a shopping cart for
types of objects as inspiration by saying, ”Let’s try to be     the elderly within 20 minutes. You must include three
more specific about what you want me to inspire with”           design inspirations from the AI in the design”. The whole
(Right side of Figure 4). The user can respond witree           task will be screen recorded. After the task, the partici-
options, ”Design Task Objects” (as our design task object       pants will fill out Creativity Support Index (CSI), which
is shopping cart, the button says ”Shopping Carts”), ”vi-       is a well known psycho metric survey, for measuring six
sually similar objects”, or ”conceptually similar objects”.     dimensions of creativity: Exploration, Expressiveness,
Visually similar objects have visual structural similarity      Immersion, Enjoyment, Results Worth Effort and Collab-
as the design task object and conceptually similar objects      oration [17] to evaluate user engagement, collaboration
have similar working mechanism or usage as the design           and immersion. After that, a retrospective think-aloud
task object. When the user clicks any of these three            will be conducted as the participants watch the screen-
buttons, the PenPal will generate a sketch accordingly.         recording video of the task to understand the rationale
                                                                behind the user interaction process and user experience.
4.5. User Finished Sketching                                    The study will end with a follow-up semi-structured in-
                                                                terview to determine in depth qualitative data about the
When the user finishes the design ideation sketching,           user engagement and overall experience with the AI.
they let the virtual agent know by clicking the ”Finish
ideation” button. The virtual agent arrives and greets the
user for completing the design ideation task by saying,         6. Discussion
”Well done! You did a great job! ”.
                                                        In the young yet fast-growing field of human-AI co-
                                                        creativity, attention is needed to design human-centered
5. Study Protocol                                       co-creative systems where users are engaged in a success-
                                                        ful collaborative experience. Interaction design where
The user experiment will take place virtually. We will users can communicate with the AI for providing user
use Google Meet to connect with the study participants. preference improves the collaborative experience and
The target sample size for the study is 50 participants user attitude towards the AI [2]. Conversing virtual
Figure 5: User Did not Like PenPal’s Sketch


agents have transitioned into services such as ecommerce,       References
leading to an increased ability for user engagement. In a
conversing interaction, users can provide their feedback        [1] N. Davis, Human-computer co-creativity: Blending
on the AI’s contribution, which provides more informa-              human and computational creativity, in: Proceed-
tion to the AI about user preferences. Conversing inter-            ings of the AAAI Conference on Artificial Intelli-
action also helps users perceive the AI as a partner rather         gence and Interactive Digital Entertainment, vol-
than a tool. A user’s confidence in an AI agent’s ability to        ume 9, 2013.
perform tasks is improved when imbuing the agent with           [2] R. Louie, A. Coenen, C. Z. Huang, M. Terry, C. J.
embodiment and social behaviors compared to the agent               Cai, Novice-ai music co-creation via ai-steering
solely depending on conversation [14]. The embodiment               tools for deep generative models, in: Proceedings
improves the user perception of an AI agent in terms of             of the 2020 CHI Conference on Human Factors in
a collaborative partner, an entity. Users also tend to trust        Computing Systems, 2020, pp. 1–13.
an AI in terms of their ability when they can see their         [3] A. Sutcliffe, Designing for user engagement: Aes-
presence. The embodiment also helps design affective AI             thetic and attractive user interfaces, Synthesis lec-
where an AI’s feelings are visible in its expression or ges-        tures on human-centered informatics 2 (2009) 1–55.
ture. As a young field and new research area, interaction       [4] O. Bown, Player responses to a live algorithm: Con-
design is rarely discussed in the existing literature despite       ceptualising computational creativity without re-
being a fundamental property of an adequate co-creative             course to human comparisons?, in: ICCC, 2015, pp.
system. An adequate interaction model dramatically im-              126–133.
proves the quality of the collaboration and engages users.      [5] Library of mixed-initiative creative interfaces,
Investigating the impact of conversing interaction and AI           http://mici.codingconduct.cc/, ???? (Accessed on
embodiment for designing effective co-creative systems              05/31/2020).
that engage users is essential.                                 [6] S. Deterding, J. Hook, R. Fiebrink, M. Gillies, J. Gow,
   We develop the prototype of Creative PenPal as an                M. Akten, G. Smith, A. Liapis, K. Compton, Mixed-
effort to explore the impact of a conversing embodied               initiative creative interfaces, in: Proceedings of
co-creative AI agent on user engagement, user percep-               the 2017 CHI Conference Extended Abstracts on
tion and overall collaborative experience. We describe              Human Factors in Computing Systems, 2017, pp.
the study design that will provide insights for designing           628–635.
effective co-creative systems that engage users and im-         [7] Y. Rogers, H. Sharp, J. Preece, Interaction design:
prove their collaborative experience with the AI agent.             beyond human-computer interaction, John Wiley
With the insights and results from the study, we will               & Sons, 2011.
improve the interaction design of Creative PenPal and           [8] P. Isola, J.-Y. Zhu, T. Zhou, A. A. Efros, Image-
implement the AI model in the improved prototype. At                to-image translation with conditional adversarial
the time of the workshop, we will have the data and in-             networks, in: Proceedings of the IEEE conference
sights in our hands from the study, revealing the impact            on computer vision and pattern recognition, 2017,
of the interaction model we used. During the workshop,              pp. 1125–1134.
we will be able to demonstrate the results and insights.        [9] C. Gutwin, S. Greenberg, M. Roseman, Workspace
     awareness in real-time distributed groupware:
     Framework, widgets, and evaluation, in: People
     and Computers XI, Springer, 1996, pp. 281–298.
[10] G. Hoffman, G. Weinberg, Interactive improvisation
     with a robotic marimba player, Autonomous Robots
     31 (2011) 133–153.
[11] N. Bryan-Kinns, F. Hamilton, Identifying mutual
     engagement, Behaviour & Information Technology
     31 (2012) 101–125.
[12] M. Yee-King, M. d’Inverno, Experience driven de-
     sign of creative systems (2016).
[13] J. Preece, H. Sharp, Y. Rogers, Interaction design:
     beyond human-computer interaction, John Wiley
     & Sons, 2015.
[14] K. Kim, L. Boelling, S. Haesler, J. Bailenson,
     G. Bruder, G. F. Welch, Does a digital assistant
     need a body? the influence of visual embodiment
     and social behavior on the perception of intelligent
     virtual agents in ar, in: 2018 IEEE International
     Symposium on Mixed and Augmented Reality (IS-
     MAR), IEEE, 2018, pp. 105–114.
[15] G. Bente, S. Rüggenberg, N. C. Krämer, Social pres-
     ence and interpersonal trust in avatar-based, col-
     laborative net-communications, in: Proceedings
     of the Seventh Annual International Workshop on
     Presence, 2004, pp. 54–61.
[16] C. Clavel, A. Cafaro, S. Campano, C. Pelachaud,
     Fostering user engagement in face-to-face human-
     agent interactions: a survey, in: Toward Robotic
     Socially Believable Behaving Systems-Volume II,
     Springer, 2016, pp. 93–120.
[17] E. Cherry, C. Latulipe, Quantifying the creativ-
     ity support of digital tools through the creativity
     support index, ACM Transactions on Computer-
     Human Interaction (TOCHI) 21 (2014) 1–25.
[18] P. Karimi, J. Rezwana, S. Siddiqui, M. L. Maher,
     N. Dehbozorgi, Creative sketching partner: an
     analysis of human-ai co-creativity, in: Proceedings
     of the 25th International Conference on Intelligent
     User Interfaces, 2020, pp. 221–230.