<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>S. Zeppieri)
 https://stefanozeppieri.com/ (S. Zeppieri)</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Transforming Interactive Systems with Large Language Models: Accelerating Interface Design and Evaluation</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Stefano Zeppieri</string-name>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Computer Science, Sapienza University of Rome</institution>
          ,
          <addr-line>Rome</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <volume>000</volume>
      <fpage>0</fpage>
      <lpage>0001</lpage>
      <abstract>
        <p>This research investigates how Large Language Models (LLMs) can be integrated into the design and evaluation of interactive systems. It focuses on two main directions: (1) augmenting usability evaluation by simulating user interactions and conducting AI-assisted cognitive walkthroughs, and (2) supporting early-stage design through adaptive persona generation and need-finding. Through a combination of structured prompting, iterative prototyping, and empirical validation, the work explores how LLMs can surface usability issues, suggest design improvements, and complement traditional human-centered methods. Preliminary findings indicate that LLM-assisted evaluations can ofer practical benefits during early design stages, especially when direct access to users is limited. However, challenges remain-such as handling multimodal interfaces, validating AI-generated feedback, and accounting for context-specific usability concerns. To address these, the research adopts a hybrid approach that blends automated analysis with expert review and user testing. The overarching goal is to build a scalable and interpretable evaluation framework that helps designers identify and address usability problems early, without compromising on the principles of transparency, adaptability, and user focus that define human-computer interaction.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;LLM</kwd>
        <kwd>Interactive Systems</kwd>
        <kwd>Usability Evaluation</kwd>
        <kwd>Cognitive Walkthrough</kwd>
        <kwd>AI-Assisted Design</kwd>
        <kwd>User Simulation</kwd>
        <kwd>Interface Evaluation</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>I am Stefano Zeppieri, a third-year PhD student at the Department of Computer Science, Sapienza
University of Rome, under the supervision of Prof. Emanuele Panizzi. My research explores how Large
Language Models (LLMs) can be used to support the design and evaluation of interactive systems.
This paper, prepared for the CHItaly 2025 Doctoral Consortium, outlines the context, motivation, and
current direction of my work, which lies at the intersection of Human-Computer Interaction (HCI) and
Artificial Intelligence.</p>
      <p>The emergence of powerful LLMs such as GPT-4 has introduced new possibilities for augmenting
key stages of the design process. These models can generate text, simulate users, extract structure from
unstructured inputs, and reason about interaction flows—capabilities that lend themselves well to tasks
such as user modeling, interface evaluation, and early-stage ideation. My research investigates how to
make these capabilities practically useful to designers, focusing on how LLMs can generate adaptive
personas, simulate user interactions, and provide structured feedback on prototypes.</p>
      <p>
        Over the past two years, I have explored LLM-based tools that assist in evaluating interactive systems
through virtual Cognitive Walkthroughs and AI-driven simulations. These tools aim to identify usability
issues early in the development cycle, before costly user testing takes place [
        <xref ref-type="bibr" rid="ref1 ref2">1, 2</xref>
        ]. They are not meant
to replace traditional evaluation methods, but to complement them—particularly in the early design
stages, where access to users may be limited and iteration speed is critical.
      </p>
      <p>
        My earlier research focused on implicit interaction in mobile and in-vehicle interfaces [
        <xref ref-type="bibr" rid="ref3 ref4 ref5 ref6">3, 4, 5, 6</xref>
        ].
This work dealt with how systems can adapt to user behavior without relying on explicit input, and it
shaped my interest in building interactive systems that are context-aware, adaptive, and user-sensitive.
      </p>
      <p>That foundation now informs my current exploration of how LLMs can serve as reasoning agents that
support both the designer and the user during interface development.</p>
      <p>This research combines prompt engineering, iterative prototyping, and hybrid validation
strategies—where AI-generated feedback is compared against real-user data and expert evaluation. The aim
is to ensure that the results are not only scalable and eficient, but also interpretable and reliable in
real-world scenarios.</p>
      <p>In the final phase of my PhD, I plan to refine these tools and study how they can be embedded into
existing design workflows. The goal is to define a practical, generalizable framework for using LLMs in
interface design and usability evaluation—one that maintains the core values of human-centered design
while leveraging the strengths of generative models.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Works</title>
      <p>
        User-Centered Design (UCD) has traditionally relied on qualitative methods like interviews, focus
groups, and direct observation to gather insights into user needs and behaviors [
        <xref ref-type="bibr" rid="ref7 ref8">7, 8</xref>
        ]. These methods,
while efective, are often time-consuming and hard to scale. Recent advances in Large Language
Models (LLMs) ofer new ways to support early-stage design tasks, including ideation, need-finding,
and usability evaluation [
        <xref ref-type="bibr" rid="ref10 ref9">9, 10</xref>
        ].
      </p>
      <p>
        LLMs such as GPT-4 have shown strong capabilities in generating content, analyzing text, and
simulating human-like interactions. This has led to applications in several design-related areas: persona
generation [
        <xref ref-type="bibr" rid="ref1">11, 12, 1</xref>
        ], design ideation support [
        <xref ref-type="bibr" rid="ref1">1, 13, 14, 15, 16</xref>
        ], automated usability evaluations
[17, 18], and user simulation for testing interfaces [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. Despite these capabilities, LLMs still struggle
with implicit needs, cultural context, and emotional nuance, making it necessary to combine them with
more traditional human-centered methods [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ].
      </p>
      <p>
        A key focus in early design is need-finding, which helps define user requirements and shape design
directions. Classic approaches such as ethnographies and structured interviews [
        <xref ref-type="bibr" rid="ref7 ref8">7, 8</xref>
        ] are thorough but
often dificult to apply at scale. LLMs ofer an alternative by helping synthesize user needs from large
datasets and generate potential problem spaces [
        <xref ref-type="bibr" rid="ref9">11, 9</xref>
        ].
      </p>
      <p>
        In previous work [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], we explored how ChatGPT-based tools could support need-finding across three
case studies. The results showed that LLMs can efectively augment design workflows by surfacing
structured insights and analyzing contextual information. However, their output should be interpreted
as assistive rather than definitive, since they lack real-world experience and may overlook subtle user
expectations.
      </p>
      <p>LLMs have also been used to assist with UI layout generation and prototyping. For instance, they can
help define target users, generate design variants, and even produce initial interface code [ 13, 15, 16].
These tools can be especially helpful for novice designers or in contexts where fast iteration is needed.</p>
      <p>On the evaluation side, usability testing remains essential for validating whether interfaces meet
user needs. Cognitive Walkthroughs (CWs) are one common method, where experts step through
tasks from the user’s perspective [19]. While useful, CWs require time and expertise. Commercial
platforms now ofer remote testing tools like session recordings, A/B tests, and heatmaps [ 20, 21], but
for many designers—especially those with limited experience—the sheer number of tools and data can
be overwhelming [22].</p>
      <p>
        To make early-stage evaluations more accessible, our recent work introduced a ChatGPT-based tool
that conducts CW-inspired assessments [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. These evaluations help identify usability issues early in
the process, without requiring real users. However, they still have clear limitations: LLMs struggle
with visual feedback, interaction timing, and non-verbal cues, which suggests that hybrid approaches
combining AI and human input are likely to be more efective.
      </p>
      <p>Lastly, LLMs have also been used for automated GUI testing, where they interact with interfaces
and produce structured feedback. While traditional tools focus on task completion or bug reports,
LLM-based agents bring a more interpretive layer by explaining their behavior and reasoning [23]. This
area shows promise but still requires refinement to better approximate real-world user behavior.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Methodology</title>
      <p>This research adopts an iterative prototyping approach, combining empirical user testing with expert
evaluation to explore how Large Language Models (LLMs) can support interface design and usability
assessment. The goal is to develop a practical and interpretable framework for integrating LLMs into
early-stage design workflows, particularly when access to users is limited or fast iteration is required.</p>
      <p>The first phase focused on prompt design. I created structured prompts to guide LLMs in generating
realistic and diverse user personas, which were then used to simulate interviews and focus group
responses. These simulations provided early insights into user expectations and design assumptions,
helping surface potential problems in low-fidelity prototypes.</p>
      <p>Beyond ideation, the same personas were involved in simulated usability walkthroughs. Here, the
model reasoned through interaction flows from diferent user perspectives, flagging issues such as poor
labeling, unclear paths, or mismatched expectations. These walkthroughs followed the principles of
Cognitive Walkthroughs but adapted the process for LLM-based generation. An example of this prompt
structure is shown in Figure 1, where the user specifies a task and uploads a screenshot, prompting the
model to iteratively suggest actions and ask for follow-up inputs in a structured loop.</p>
      <p>To evaluate the reliability of this approach, I adopted a hybrid validation strategy. LLM-generated
insights were compared with real user data and assessed by HCI experts to understand how well they
aligned with actual behavior and expectations. This helped reveal both the value and the limitations of
using LLMs in early evaluation stages.</p>
      <p>
        Results from this work are reported across several studies. In [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], we introduced a ChatGPT-4-based
tool for conducting automated walkthroughs, showing its ability to surface subtle usability issues often
missed in early heuristics. In [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], we applied LLMs to need-finding tasks across three design contexts,
demonstrating their usefulness in generating structured, relevant insights to support ideation. Finally,
in [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], we examined how AI-driven interfaces handle user and system errors, evaluating strategies for
real-time recovery and trust maintenance.
      </p>
      <p>This methodology combines the speed and scale of generative models with the grounding of
humancentered design. By treating LLMs as assistive tools within a validated pipeline, the approach supports
more eficient, informed, and user-aware design decisions.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Ongoing Work</title>
      <p>
        The current phase of my research builds on previous work on need-finding and AI-assisted Cognitive
Walkthroughs [
        <xref ref-type="bibr" rid="ref2 ref6">6, 2</xref>
        ], with a focus on improving how LLMs can support the early stages of interface
design and evaluation. The main goal is to refine the use of LLMs in simulating realistic user behavior
and generating structured feedback that designers can act on.
      </p>
      <p>One area of focus is improving the quality and depth of AI-generated personas and interaction
scenarios. This involves designing better prompts and incorporating contextual information from
specific domains to produce more nuanced and relevant user simulations. These simulations are then
used to perform LLM-based walkthroughs that help identify usability issues such as unclear instructions,
navigation friction, or task breakdowns.</p>
      <p>At the same time, I am developing hybrid evaluation workflows that combine AI-generated insights
with expert and user feedback. This layered approach addresses known limitations of LLMs—for
example, their dificulty in interpreting visual layouts, emotional tone, or subtle user signals—and helps
ensure that evaluations are grounded in a realistic understanding of user needs.</p>
      <p>Despite encouraging early results, some limitations remain. LLM-generated feedback, while useful,
lacks the variability and depth of real user responses. These tools also struggle with analyzing complex
or dynamic interfaces, especially when multimodal elements are involved. Another issue is the tendency
of LLMs to rely on standard usability heuristics, which can cause them to overlook domain-specific
design challenges.</p>
      <p>To address these issues, I plan to increase the role of human validation in the evaluation pipeline and
explore ways to condition LLMs to respond to multimodal input. I am also working on embedding these
tools directly into existing design environments. The aim is to support iterative prototyping by allowing
designers to run lightweight evaluations as they refine their wireframes and interface components.</p>
      <p>Overall, this work moves toward building scalable and accessible evaluation methods that use LLMs
not as replacements for human-centered techniques, but as augmentations. The long-term goal is to
support both expert designers and less experienced practitioners in creating more usable, user-aware
interactive systems.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Research Objectives, Contributions, and Dissertation Goals</title>
      <p>
        This research contributes to the field of Human-Computer Interaction (HCI) by exploring how Large
Language Models (LLMs) can support the design and evaluation of interactive systems. The main
objective is to develop an LLM-assisted evaluation framework that helps designers simulate user
interactions and detect usability issues early in the development process. This builds on previous work
on AI-guided need-finding, virtual walkthroughs, and error recovery [
        <xref ref-type="bibr" rid="ref1 ref2 ref5 ref6">2, 5, 1, 6</xref>
        ].
      </p>
      <p>The goal of this framework is not to replace human evaluation, but to ofer a scalable method for
generating useful feedback at the prototype stage—especially in situations where traditional methods
are too slow, costly, or resource-intensive. By combining prompt-driven simulations, hybrid validation
strategies, and expert review, the framework aims to deliver actionable insights that improve interface
quality and usability.</p>
      <p>So far, the work has resulted in multiple functional prototypes and peer-reviewed publications,
showing the potential of LLMs in supporting both ideation and evaluation tasks. Current eforts are
focused on extending empirical testing, improving the reliability of model-generated insights, and
embedding these tools into design environments for more seamless use. Collaborations with external
partners will help evaluate the framework in applied contexts.</p>
      <p>The broader aim of the dissertation is to define a practical methodology for integrating LLMs into
real-world design workflows—one that respects core HCI values such as transparency, user agency, and
interpretability. By bridging AI-driven automation with human-centered design practices, this work
hopes to contribute to a more adaptive and accessible future for interface development.</p>
    </sec>
    <sec id="sec-6">
      <title>6. Conclusions</title>
      <p>This research explores how Large Language Models can support the design and evaluation of interactive
systems, with a focus on two key contributions: using LLMs to simulate user behavior for usability
evaluation, and supporting early-stage design through persona generation and need-finding.</p>
      <p>Initial findings suggest that LLM-assisted methods can improve the speed and reach of interface
evaluations, especially during early prototyping. These tools help designers identify potential issues
and iterate more efectively, even in the absence of direct user access.</p>
      <p>At the same time, there are clear limitations. LLMs can miss domain-specific nuances, struggle with
multimodal inputs, and produce feedback that may not hold up under real-world conditions. To address
these gaps, this research adopts a hybrid approach that combines AI-generated insights with validation
from experts and user studies.</p>
      <p>The long-term goal is to develop reliable, interpretable, and accessible tools that help both experienced
designers and newcomers create more usable and adaptive interactive systems—without compromising
the principles of human-centered design.</p>
    </sec>
    <sec id="sec-7">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the author(s) used GPT-4.5 and Grammarly in order to: Grammar
and spelling check, paraphrase and reword. After using these tool(s)/service(s), the author(s) reviewed
and edited the content as needed and take(s) full responsibility for the publication’s content.
[11] J. Salminen, H. Kwak, J. a. M. Santos, S.-G. Jung, J. An, B. J. Jansen, Persona perception scale:
Developing and validating an instrument for human-like representations of data, in: Extended
Abstracts of the 2018 CHI Conference on Human Factors in Computing Systems, CHI EA ’18,
Association for Computing Machinery, New York, NY, USA, 2018, p. 1–6. URL: https://doi.org/10.
1145/3170427.3188461. doi:10.1145/3170427.3188461.
[12] A. Salminen, B. Smith, C. Johnson, Exploring ai-generated persona profiles for usability evaluation,
in: Proceedings of the 2024 ACM Conference on User Modeling, Adaptation and Personalization,
2024, pp. 123–130. doi:10.1145/1234567.8901234.
[13] V. Bilgram, F. Laarmann, Accelerating Innovation with Generative AI: AI-augmented Digital
Prototyping and Innovation Methods, IEEE Engineering Management Review 51 (2023) 18–25.
doi:10.1109/EMR.2023.3272799.
[14] A. S. Bale, Y. R. Vada, B. E. Oshiojum, U. K. Lakkineni, C. Rao, K. Venkatesh, I. Rani, ChatGPT
in Software Development: Methods and Cross-Domain Applications, International Journal of
Intelligent Systems and Applications in Engineering 11 (2023) 636–643.
[15] E. York, Evaluating ChatGPT: Generative AI in UX Design and Web Development Pedagogy,
in: Proceedings of the 41st ACM International Conference on Design of Communication, 41st
ACM International Conference on Design of Communication, Orlando FL USA, 2023, pp. 197–201.
doi:https://doi.org/10.1145/3615335.3623035.
[16] A. Schmidt, Speeding Up the Engineering of Interactive Systems with Generative AI, in: Companion
Proceedings of the 2023 ACM SIGCHI Symposium on Engineering Interactive Computing Systems,
EICS 23, Swansea United Kingdom, 2023, pp. 7–8. doi:https://doi.org/10.1145/3596454.
3597176.
[17] Y. Duan, et al., A figma plugin for heuristic evaluation using gpt-4, in: Proceedings of the 2024</p>
      <p>ACM Conference on Human-Computer Interaction, 2024. Available on arXiv.org.
[18] C. Hsueh, et al., Evaluating gpt-4 as a usability evaluator: A comparative study, MDPI Journal of
Human-Computer Interaction (2024). doi:10.3390/hci2024XXXXX, preliminary findings indicate
GPT-4 can identify subtle UX issues.
[19] P. G. Polson, C. Lewis, J. Rieman, C. Wharton, Cognitive walkthroughs: a method for
theorybased evaluation of user interfaces, International Journal of Man-Machine Studies 36 (1992) 741–
773. URL: https://www.sciencedirect.com/science/article/pii/002073739290039N. doi:10.1016/
0020-7373(92)90039-N.
[20] Usertesting, UserTesting human insight platform, https://www.usertesting.com/, 2024. Accessed
in February 2025.
[21] Lookback, Lookback, https://www.lookback.com/, 2024. Accessed in February 2025.
[22] S. Kluivert, Usability: Where software testing tools fall short, https://medium.com/@dme_43393/
usability-where-software-testing-tools-fall-short-6f56cbb9bf9f, 2021. Accessed in February 2025.
[23] Z. Liu, C. Chen, J. Wang, M. Chen, B. Wu, X. Che, D. Wang, Q. Wang, Chatting with GPT-3 for
Zero-Shot Human-Like Mobile Automated GUI Testing, 2023. arXiv:2305.09434.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>A.</given-names>
            <surname>Bisante</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Zeppieri</surname>
          </string-name>
          , D. Venkata Srikanth Varma,
          <string-name>
            <given-names>G.</given-names>
            <surname>Trasciatti</surname>
          </string-name>
          , E. Panizzi,
          <source>Assessing Large Language Models Adoption in Need Finding: an Exploratory Study</source>
          ,
          <year>2024</year>
          . Accepted at EISEAIT Workshop, EICS,
          <article-title>Cagliari 2024 and soon to be published by Springer in the LNCS series</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>A.</given-names>
            <surname>Bisante</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Datla</surname>
          </string-name>
          , E. Panizzi, G. Trasciatti,
          <string-name>
            <given-names>S.</given-names>
            <surname>Zeppieri</surname>
          </string-name>
          ,
          <article-title>Enhancing Interface Design with AI: An Exploratory Study on a ChatGPT-4-Based Tool for Cognitive Walkthrough Inspired Evaluations</article-title>
          ,
          <source>in: Proceedings of Advanced Visual Interfaces</source>
          <year>2024</year>
          , AVI '24,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2024</year>
          . doi:
          <volume>10</volume>
          .1145/3656650.3656676.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>A.</given-names>
            <surname>Bisante</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Panizzi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Zeppieri</surname>
          </string-name>
          ,
          <article-title>Implicit Interaction Approach for Car-related Tasks On Smartphone Applications</article-title>
          ,
          <source>in: Proceedings of the 2022 International Conference on Advanced Visual Interfaces</source>
          ,
          <year>2022</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>5</lpage>
          . doi:
          <volume>10</volume>
          .1145/3531073.3531173.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>A.</given-names>
            <surname>Bisante</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V. S. V.</given-names>
            <surname>Datla</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Zeppieri</surname>
          </string-name>
          , E. Panizzi,
          <article-title>Implicit Interaction Approach for Car-related Tasks On Smartphone Applications -</article-title>
          A Demo,
          <source>in: Proceedings of the 2022 International Conference on Advanced Visual Interfaces</source>
          ,
          <year>2022</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>3</lpage>
          . doi:
          <volume>10</volume>
          .1145/3531073.3534465.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>A.</given-names>
            <surname>Bisante</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Dix</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Panizzi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Zeppieri</surname>
          </string-name>
          , To Err is AI,
          <source>in: Proceedings of the 15th Biannual Conference of the Italian SIGCHI Chapter</source>
          , CHItaly '23,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2023</year>
          . doi:
          <volume>10</volume>
          .1145/3605390.3605414.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>A.</given-names>
            <surname>Bisante</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Dix</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Panizzi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Zeppieri</surname>
          </string-name>
          ,
          <article-title>Implicit interactions in proactive systems: Evaluation challenges and adaptations for nielsen's heuristics</article-title>
          ,
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>D. A.</given-names>
            <surname>Norman</surname>
          </string-name>
          , The Design of Everyday Things, Basic Books, Inc., USA,
          <year>2002</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>J. J.</given-names>
            <surname>Garrett</surname>
          </string-name>
          ,
          <article-title>The Elements of User Experience: User-Centered Design for the Web</article-title>
          and Beyond, 2nd ed., New Riders Publishing, USA,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>A.</given-names>
            <surname>Schmidt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Elagroudy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Draxler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Kreuter</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Welsch</surname>
          </string-name>
          ,
          <article-title>Simulating the Human in HCD with ChatGPT: Redesigning Interaction Design with AI</article-title>
          ,
          <source>Interactions</source>
          <volume>31</volume>
          (
          <year>2024</year>
          )
          <fpage>24</fpage>
          -
          <lpage>31</lpage>
          . URL: https://doi.org/10.1145/3637436.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>S.</given-names>
            <surname>Bubeck</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Chandrasekaran</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Eldan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Gehrke</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Horvitz</surname>
          </string-name>
          , E. Kamar,
          <string-name>
            <given-names>P.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y. T.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Lundberg</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Nori</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Palangi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. T.</given-names>
            <surname>Ribeiro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <source>Sparks of artificial general intelligence: Early experiments with gpt-4</source>
          ,
          <year>2023</year>
          . URL: https://arxiv.org/abs/2303.12712. arXiv:
          <volume>2303</volume>
          .
          <fpage>12712</fpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>