<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <issn pub-type="ppub">1613-0073</issn>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>multi-phase pilot study on multimodal evaluation in cognitive tests</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Alejandro Enriquez</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Juan Camilo Méndez Flórez</string-name>
          <email>juancamilo.mendez.florez@usc.es</email>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Nelly Condori-Fernández</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Alejandro Catala</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Workshop</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Departamento de Electrónica e Computación, Universidade de Santiago de Compostela</institution>
          ,
          <addr-line>15782 Santiago de Compostela</addr-line>
          ,
          <country country="ES">Spain</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>RadAmbiente</institution>
          ,
          <addr-line>Cuenca</addr-line>
          ,
          <country country="EC">Ecuador</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>de Compostela</institution>
          ,
          <country country="ES">Spain</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Cognitive impairment is an increasingly prevalent issue, particularly with the ageing population. Brief cognitive screening tests play a key role in early detection, and their progressive digitisation raises new methodological challenges. One such test, the SAGE (Self-Administered Gerocognitive Exam), which assesses areas such as orientation, language, and memory, has proven useful for the early detection of cognitive impairment. This paper presents a pilot study designed to compare the traditional and digital formats of a cognitive test based on a multimodal protocol. The protocol integrates objective performance scores, subjective assessments of user experience (UX) and mental workload, and emotional response analysis conducted in two sequential phases: the ifrst using facial expression analysis, and the second focusing exclusively on stress via electrodermal activity (EDA) sensors. The second phase was conducted in care home environments, where ethical constraints prohibited the use of cameras, highlighting the protocol's adaptability to real-world sensitive contexts. This paper contributes to understanding how test format influences not only cognitive outcomes but also UX and perceived mental workload, providing a replicable protocol for evaluating the digital transition of cognitive assessment tools.</p>
      </abstract>
      <kwd-group>
        <kwd>cognitive impairment</kwd>
        <kwd>digital SAGE test</kwd>
        <kwd>pilot study</kwd>
        <kwd>UX</kwd>
        <kwd>emotional data</kwd>
        <kwd>electrodermal activity</kwd>
        <kwd>workload</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Cognitive impairment is a growing challenge for global public health, driven by an ageing population
and aggravated by factors such as unwanted loneliness among older adults [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. In Spain, this issue is
particularly pressing: according to Alzheimer Europe (2019), 1.83% of the population already shows
signs of cognitive impairment, a figure projected to more than double by 2050, reaching 3.99% [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. In
this context, the early detection of mild cognitive impairment (MCI) has become a key priority.
      </p>
      <p>
        Among the tools available for this purpose, the Self-Administered Gerocognitive Exam (SAGE) [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]
stands out. This clinically validated, self-administered test enables the early detection of MCI by
assessing cognitive domains such as orientation, language, memory, executive function, and visuoconstructive
skills [
        <xref ref-type="bibr" rid="ref4 ref5">4, 5</xref>
        ]. Its ease of use makes it a practical, remote, and scalable option.
      </p>
      <p>
        The digitisation of the SAGE test has been motivated by the need to increase its accessibility,
particularly for vulnerable populations such as older adults with mobility limitations, those living in rural areas,
or individuals experiencing social isolation [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. While mobile or tablet-based versions ofer potential
benefits such as self-monitoring and immediate feedback, the shift from paper-based to digital formats
introduces questions about usability, user experience, and cognitive performance comparability.
      </p>
      <p>
        This paper presents a pilot study aimed at comparing the traditional and digital formats of a cognitive
test, taking SAGE as a representative case. The digital format used in this study corresponds to the
version developed in [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. Our experimental protocol integrates objective performance scores, subjective
      </p>
      <p>CEUR</p>
      <p>ceur-ws.org
assessments of user experience and mental workload, and emotional response analysis conducted in
two sequential phases that correspond to diferent settings. In the first phase, emotional data were
gathered via facial expression analysis in a lab setting. In the second phase, conducted in care home
environments, emotional responses were monitored through electrodermal activity (EDA) sensors, due
to restrictions on video recordings in such settings.</p>
      <p>This pilot study highlights the adaptability of a multimodal evaluation to ethically sensitive
realworld contexts, where traditional data collection methods may be restricted, ofering insights for the
development of digital cognitive testing approaches that balance methodological rigour with user
acceptability.</p>
      <p>The remainder of this paper is organised as follows: Section 2 provides background and reviews
related work on the digitisation of cognitive tests. Section 3 presents the methodology, including the
study’s objective, variables and experimental procedure. Section 4 reports the initial results. Finally,
Section 5 discusses some key aspects to be considered in future extended formal evaluations.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Background</title>
      <sec id="sec-2-1">
        <title>2.1. Self-administered gerocognitive examination (SAGE) test</title>
        <p>
          Several tools have been developed considering the importance of early detection and prevention.
For this study, the Self-Administered Gerocognitive Examination (SAGE) test will be brought into
focus. Its reliability and validity as a tool for early detection of dementia and Alzheimer’s disease are
well-documented through several studies [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ]. This test evaluates diferent cognitive areas, such as:
1. Orientation: This domain assesses awareness of time, date, and place, which are essential for
daily functioning. Deficits may indicate early cognitive impairment, common in dementia.
2. Naming: Assesses the ability to accurately identify and name objects, involving the involvement
of the brain’s language centers.
3. Similarities: Assesses reasoning and abstract thinking by comparing objects or concepts.
4. Calculation: It evaluates cognitive skills related to numerical processing, logical thinking, and
mathematical problem-solving.
5. Memory: This domain assesses the capacity to encode, store, and retrieve new information.
6. Construction: Evaluates visuospatial skills and the ability to plan and execute movements for
the purpose of copying or constructing figures.
7. Verbal Fluency: This area evaluates the ability to generate words rapidly and eficiently under
specific constraints, measuring executive function, lexical retrieval, and language fluency.
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Related work</title>
        <p>
          This section critically discusses the similarities and diferences between the current study and related
literature in early detection and prevention of Alzheimer’s and dementia. In terms of the validity of the
SAGE test, [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ] evaluated 665 patients, comparing the results of the SAGE with the Mini-Mental State
Examination (MMSE). The findings revealed that the SAGE detects dementia six months earlier than
the MMSE and, being self-administered, removes implementation barriers associated with the need for
medical personnel. Additionally, it highlights that repeated SAGE scores function as a reliable cognitive
biomarker for monitoring the progression of impairment.
        </p>
        <p>
          Other experiments have validated the applicability of the test in broader contexts, demonstrating
that the SAGE test is feasible, practical, reliable, and efective in various settings and among diferent
population groups [
          <xref ref-type="bibr" rid="ref4 ref7">7, 4</xref>
          ].
        </p>
        <p>
          Finally, Scharre et al. [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] found a strong correlation between the paper SAGE scores and its paid
digital version, suggesting that the digital format is equally valid. It does not provide details about user
experience and usability design issues.
        </p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Methodology</title>
      <p>This section describes the methodological approach of the pilot study. Given its exploratory nature,
the study adopts a multi-phase research design in which each phase addresses complementary aspects
of the overall goal through specific data collection techniques and settings. This design allowed the
protocol to adapt to evolving constraints in both controlled and real-world environments.</p>
      <sec id="sec-3-1">
        <title>3.1. Goal and research questions of the study</title>
        <p>
          The main objective of this study is structured using the Goal–Question–Metric template [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ]:
Analyze diferences associated with the test format (traditional vs. digital) for the purpose
of comparing cognitive performance, perceived user experience, mental workload, and
emotional responses with respect to the outcomes observed through a multimodal evaluation
approach from the point of view of researchers and practitioners interested in digital
cognitive screening in the context of a pilot study conducted in two phases, including
real-world ethically sensitive settings.
        </p>
        <p>The study is guided by the following research questions and related variables. The dependent
variables measured are detailed in Table 1.</p>
        <p>
          • RQ1. How diferent is the cognitive performance between the traditional and digital
formats of the cognitive test?
This question explores whether the format of the cognitive test influences participants’
performance, which is crucial for assessing the comparability of both formats in the early detection
of mild cognitive impairment (MCI). Cognitive performance is measured using the test scores
(range: 0–22) obtained in both the traditional and digital formats of the SAGE test.
• RQ2. How does the test format (traditional vs. digital) afect the perceived workload?
This question aims to explore whether participants perceive diferent levels of workload
depending on the test format. In the study, perceived workload was measured using the NASA
TLX questionnaire [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ], which captures participants’ subjective impressions of task demands.
The instrument uses a scale to rate six dimensions: mental demand, physical demand, temporal
demand, performance, efort, and frustration.
• RQ3. How do participants’ emotional responses difer depending on the test format?
This question examines whether participants’ emotional responses vary according to the test
format. Participants’ emotional responses during the completion of the SAGE test were gathered.
Emotional data related to frustration (from 0% to 100%, indicating the proportion of time facial
expressions related to that emotion were detected during the test), valence (from -1 to 1, indicating
negative to positive emotions), and arousal (from -1 to 1, indicating emotional activation levels)
were obtained through facial analysis using Kopernica Human, an artificial intelligence tool
developed by Neurologyca1. Additionally, stress was assessed using physiological responses from
the EDA sensor, recorded with the EmbracePlus smartwatch device by Empatica2. The data were
processed by an automatic stress detector that outputs discrete stress levels from 1 to 5, where
levels 4 and 5 indicate stress [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ].
        </p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Experimental procedure</title>
        <p>
          The study was structured in two phases to evaluate cognitive performance, mental workload, and
emotional responses elicited by the SAGE test in both its traditional (paper-based) and digital formats.
In both phases, the digital format of the SAGE developed in [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] was used, available as an APK for Android
1Neurologyca, Kopernica: Real-time emotional intelligence, 2025. URL: https://www.kopernica.ai/
2Empatica, EmbracePlus, URL: https://www.empatica.com/en-eu/embraceplus/
devices and specifically designed for use on touchscreen tablets. Figure 1 shows the experimental
procedure and instruments used across both study phases.
        </p>
        <p>The implementation of each phase was subject to ethical approvals and institutional permissions.
Participants in both phases were required to meet the following inclusion criteria: no prior formal
diagnosis of cognitive impairment, and provision of informed consent.</p>
        <sec id="sec-3-2-1">
          <title>3.2.1. Phase 1: lab setting</title>
          <p>In Phase 1, conducted in a laboratory setting, the SAGE test was administered in both formats under
controlled conditions, with a main focus on the evaluation of the digital version. This version was
designed based on usability principles for older adults and aimed at reducing cognitive load, minimising
errors, and ensuring an accessible experience through large buttons and typography, clear instructions,
linear navigation, and a minimalist design with emotionally positive colours (i.e., Blue, White and
Yellow).</p>
          <p>Participants were selected through convenience sampling, based on their availability, proximity and
voluntary participation. To assess emotional responses during the administration of the digital test,
participants’ facial expressions were recorded using a camera and OBS Studio, a free video recording
software configured specifically to capture close-up video in MP4 format. The videos were used as
input in the Kopernica tool, which applies facial emotion analysis algorithms and generates CSV files
with the corresponding results for interest, frustration, valence, and arousal percentages.</p>
          <p>Immediately after the test, participants responded to the NASA TLX Workload questionnaire to
assess perceived workload. The order of test format presentation was randomized across participants,
and a one-week gap was maintained between the sessions for each format.</p>
        </sec>
        <sec id="sec-3-2-2">
          <title>3.2.2. Phase 2: home care setting</title>
          <p>In Phase 2, carried out in a home care setting, participants were older adults residing in a local nursing
home, which allowed for the inclusion of a more diverse population and testing under uncontrolled
conditions, in contrast to Phase 1. Each participant completed both the digital and traditional versions
of the test, with at least a 7-day interval between each administration to avoid short-term memory
efects. In addition to cognitive scores, physiological data were collected during both tests to evaluate
user experience.</p>
          <p>For physiological data collection, the EmbracePlus smartwatch, a wearable device that records
realtime electrodermal activity data through integrated electrodes that must be in contact with the skin
to capture accurate information, was used. It operates with a sampling frequency between 1 and
4 Hz, within a range of 0.01 to 100 microsiemens (μS). To initiate monitoring, each participant was
previously assigned a unique credential through Empatica’s Care Lab Portal, which generates a QR code
used to pair the watch with a mobile phone via Bluetooth. Once paired, the device was placed on the
non-dominant wrist of the participant. Raw sensor data collected from the wearable device were stored
in AVRO format, a compact binary format for eficient data storage and fast access, and subsequently
converted to CSV using a Python script provided by Empatica for further analysis. The NASA TLX
Workload questionnaire was also administered after each format.</p>
          <p>All participants provided informed consent prior to participation, in accordance with the guidelines
approved by the relevant ethics committees. In the home care setting, consent from legal guardians or
family members, when applicable, was managed directly by the facility staf. To protect participants’
privacy, video recording was restricted in Phase 2, and only non-intrusive physiological data were
collected using a wearable device. Data from the EmbracePlus smartwatch was managed by Empatica
under a research license valid for the duration of the project. For video-based emotional analysis in
Phase 1, only numerical output from the tool was stored, and no raw video footage was retained. In line
with the agreement established with the care facility, the research team also provided both individual
and group-level summaries of test results, along with tailored cognitive health recommendations.</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Initial results</title>
      <sec id="sec-4-1">
        <title>4.1. Phase 1: lab setting</title>
        <p>Five older adults (three individuals aged between 60 and 80 and two individuals over 80) completed the
SAGE test in both formats.</p>
        <p>Table 2-Phase 1 shows the SAGE test scores for participants in the first phase, comparing both
the digital and traditional formats and including the diference in performance per participant. The
scores showed high consistency between the traditional and digital formats for most participants. For
example, P2, P3, and P4 obtained identical scores in both formats (19, 11, and 17, respectively), while the
remaining two (P1 and P5) showed a small diference of two points in favour of the traditional version.
These results, under controlled conditions, suggest that cognitive performance is generally consistent
across formats, although limited digital experience may slightly influence outcomes in favour of the
traditional version, likely due to lower familiarity with mobile devices, which is related to age—as in
the case of P1, who is 91 years old.</p>
        <p>Regarding mental workload (see values in Table 3-Phase 1), the NASA TLX Workload data for the
digital version showed significant variations among participants. The average workload ranged from
-6.0 to 7.2, with P4 reporting the highest workload (7.2) and P2 the lowest (-6.0). Participant P4 indicated
a notably demand with the digital format, despite obtaining identical scores in both formats (17). In
contrast, P2 reported negative values in most dimensions, suggesting a comfortable and undemanding
experience.</p>
        <p>In terms of emotional responses, the overall analysis of valence, arousal, and frustration (processed
through Kopernica) in the digital version revealed that most participants are located in the lower-right
quadrant of the afective circumplex, corresponding to a ”Calm” emotional state, characterized by
positive valence and low arousal (see Figure 2). This suggests that these participants felt generally
comfortable and relaxed while completing the test. An exception was observed in the case of Participant
P2, who exhibited slightly negative valence and higher arousal, placing between ”Sad” and ”Upset”
zones. According to Figure 3-a, this participant also reported the highest frustration percentage (34%).</p>
        <p>These results suggest the digital format is comparable in performance but poses challenges for
some participants, reflected in higher frustration and workload. The prevailing calm indicates good
acceptance, but cases like P2 highlight the need for more intuitive interfaces to reduce frustration.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Phase 2: home care setting</title>
        <p>Seven older adults (five women, two men; ages: 90, 76, 75, 80, 70, 90, 80, respectively) completed the
SAGE test in both formats in a more realistic setting. Table 2-Phase 2 shows the SAGE test scores for
participants in this phase. The group showed a general trend of improvement in the digital format,
where average scores increased from 13 (traditional) to 16 (digital). Regarding mental workload (see
values in Table 3-Phase 2), some diferences between formats were also observed. In the digital format,
average workload ranged from -7.7 to 4.5, with U6 showing that the task was demanding despite
improvement in performance (14 to 18). U4, with the lowest average (-7.7), reported negative values
across all dimensions, indicating a very comfortable experience. In the traditional format, averages
were more homogeneous, ranging from -3.7 to 1.0. However, the overall trend indicates that while the
digital format improved performance, it may introduce additional demands for some.</p>
        <p>The Stress levels graphs, (see Figure 3-b), revealed that stress was generally lower in the digital
format, with lower average peaks compared to the traditional version. These results indicate that in
this group of individuals, overall cognitive performance improved in the digital format and stress was
reduced, but variations in workload were introduced.</p>
        <p>Table 4 summarises stress-related measures for each participant and test format in Phase 2. The table
includes the percentage of time spent in high stress levels (≥4), average stress level, standard deviation,
and the number of detected stress episodes based on EDA signals. When comparing formats, most
participants exhibited slightly higher stress levels and more frequent stress episodes in the traditional
version than in the digital one. This pattern is particularly evident in the percentage of time spent at
stress levels 4 and 5, which tended to be greater under the paper-based condition. These diferences may
be partially explained by the improvements made to the digital prototype following usability feedback
in Phase 1. Additionally, the fixed font size and layout of the paper-based test could have introduced
visual strain for some participants.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Discussion and conclusions</title>
      <p>This pilot study tested a multimodal protocol in two distinct settings to compare cognitive performance,
workload, and emotional responses across digital and traditional formats of a cognitive test. Several
aspects of the protocol proved efective across both phases. Paper-based instruments—used for the
traditional cognitive test and workload measurement—were reliable and well accepted in both contexts.
The informed consent process was also simple, with most participants signing voluntarily without
raising concerns.</p>
      <p>In both settings, certain test items required verbal clarification during administration, and a
thinkaloud protocol could have enriched feedback beyond written responses. Most participants completed
both test formats, with two exceptions in Phase 2 due to contextual constraints.</p>
      <p>Scheduling was more consistent in Phase 1, with randomised order and a one-week interval between
formats. In Phase 2, scheduling varied due to participants’ availability or health issues, extending in
some cases to three weeks or resulting in partial data.</p>
      <p>Distinct challenges emerged in each phase. In Phase 1, one of the main dificulties involved managing
and analysing video recordings used for emotion detection via Kopernica. The videos had to be
processed ofline, and in several cases, facial detection accuracy was compromised due to posture.
Therefore, emotion recognition based on facial expressions is only available for the digital format. A few
participants also expressed concern about receiving feedback on their cognitive performance, which
may reflect a general sensitivity to test results in this demographic.</p>
      <p>In Phase 2, ethical restrictions within the home care setting strongly influenced data collection. Video
recording was not allowed at all; instead, physiological data were gathered through the EmbracePlus
wearable as an acceptable alternative. Although this method limited emotional data granularity,
restricting it mainly to EDA-based stress levels, it was well tolerated. Figure 3-b shows that stress in
digital format starts high but decreases and stabilises, while on paper it gradually increases. Table 4
confirms that most participants experienced more episodes and longer durations of elevated stress
on paper, possibly due to usability improvements in the digital version and visual discomfort with
the traditional format. Participants often reported being unaware of the device. Thus, using
nonobstructive acquisition methods could be a preferable alternative in settings where administrators/users
are reluctant to camera-based methods.</p>
      <p>Participant recruitment in Phase 2 was coordinated by the facility, which limited researchers’ ability
to select or balance participant profiles. Still, the core testing instruments were applied consistently
across both phases, with variation only in the method of emotion measurement. These contextual and
methodological diferences were documented and considered in interpreting results.</p>
      <p>Overall, this paper contributes a protocol for conducting a pilot involving two diferent types of
settings, which aimed at gauging diferences in cognitive performance and workload influenced by the
implementation format of the SAGE tests. The pilot reported initial results, providing some encouraging
preliminary results on whether the digital version may pose or not a serious entry barrier before running
a more formal experiment, It also illustrated how multimodal measurements concerned with emotional
responses can be used to complement the performance and workload measurements.</p>
      <p>This pilot study highlights the importance of planning for granular physiological data collection,
balanced test scheduling, and adaptive digital design when working with older adults. Despite contextual
constraints, the proposed multimodal evaluation protocol proved adaptable to both controlled and
real-world settings. In this regard, the consistent use of the core instruments across both phases, despite
diferences in emotional measurement, supports the protocol’s scalability. However, ethical aspects
such as informed consent, data privacy, and participant vulnerability must be adapted to each setting.
Addressing them early is key to ensuring ethical and contextual suitability at scale.</p>
      <p>These observations support the feasibility of replicating this multi-phase approach in similar studies
evaluating digital transitions in cognitive testing. The protocol ofers a comprehensive framework to
examine the impact of test format in elderly populations.</p>
    </sec>
    <sec id="sec-6">
      <title>Acknowledgements</title>
      <p>Authors acknowledge the support of the Galician Ministry of Culture, Education, Professional Training
and University (grants ED431G2023/04 and ED431C2022/19). Supported also by Interreg VI-A
SpainPortugal Program (POCTEP) 2021-2027 with grant 0144_TRANSFIRESAUDE_1_E, CNS2024-154915 by
MCIN/AEI, the Erasmus Mundus Joint Master Degree program SE4GD-619839, and the ERDF.</p>
    </sec>
    <sec id="sec-7">
      <title>Declaration on generative AI</title>
      <p>During the preparation of this work, the author(s) used Grammarly for Grammar and spelling checks.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>L.</given-names>
            <surname>Qiao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Tang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Min</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Yin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <article-title>Association between loneliness and dementia risk: A slr and meta-analysis of cohort studies, Front</article-title>
          . Hum. Neurosci.
          <volume>16</volume>
          (
          <year>2022</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>Alzheimer</given-names>
            <surname>Europe</surname>
          </string-name>
          ,
          <source>Dementia in Europe Yearbook</source>
          <year>2019</year>
          :
          <article-title>Estimating the prevalence of dementia in Europe</article-title>
          , Alzheimer Europe,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>Wexner</given-names>
            <surname>Medical</surname>
          </string-name>
          <string-name>
            <surname>Center</surname>
          </string-name>
          ,
          <article-title>Sage: The self-administered gerocognitive exam</article-title>
          ,
          <year>2023</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>D. W.</given-names>
            <surname>Scharre</surname>
          </string-name>
          ,
          <string-name>
            <surname>S.-I. Chang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. A.</given-names>
            <surname>Murden</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Lamb</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. Q.</given-names>
            <surname>Beversdorf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Kataki</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H. N.</given-names>
            <surname>Nagaraja</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. A.</given-names>
            <surname>Bornstein</surname>
          </string-name>
          ,
          <article-title>Self-administered gerocognitive examination (sage): A brief cognitive assessment instrument for mild cognitive impairment (mci) and early dementia</article-title>
          ,
          <source>Alzheimer Disease Associated Disorders</source>
          <volume>24</volume>
          (
          <year>2010</year>
          )
          <fpage>64</fpage>
          -
          <lpage>71</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>D. W.</given-names>
            <surname>Scharre</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.-I.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H. N.</given-names>
            <surname>Nagaraja</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N. C.</given-names>
            <surname>Wheeler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Kataki</surname>
          </string-name>
          ,
          <article-title>Self-administered gerocognitive examination: longitudinal cohort testing for the early detection of dementia conversion</article-title>
          ,
          <source>Alzheimers Res. Ther</source>
          .
          <volume>13</volume>
          (
          <year>2021</year>
          )
          <fpage>192</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>A.</given-names>
            <surname>Enriquez</surname>
          </string-name>
          <string-name>
            <surname>Mancheno</surname>
          </string-name>
          ,
          <article-title>Digitization of the SAGE test: building a scalable solution for early cognitive impairment detection</article-title>
          ,
          <source>Master's thesis</source>
          , Lappeenranta-Lahti University of Technology LUT, Finland,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>D. W.</given-names>
            <surname>Scharre</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.-I.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H. N.</given-names>
            <surname>Nagaraja</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Yager-Schweller</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. A.</given-names>
            <surname>Murden</surname>
          </string-name>
          ,
          <article-title>Community cognitive screening using the self-administered gerocognitive examination (sage)</article-title>
          ,
          <source>J. Neuropsychiatry Clin. Neurosci</source>
          .
          <volume>26</volume>
          (
          <year>2014</year>
          )
          <fpage>369</fpage>
          -
          <lpage>375</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>D. W.</given-names>
            <surname>Scharre</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. I.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H. N.</given-names>
            <surname>Nagaraja</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N. E.</given-names>
            <surname>Vrettos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. A.</given-names>
            <surname>Bornstein</surname>
          </string-name>
          ,
          <article-title>Digitally translated self-administered gerocognitive examination (esage): relationship with its validated paper version, neuropsychological evaluations, and clinical assessments</article-title>
          ,
          <source>Alzheimers Res. Ther</source>
          .
          <volume>9</volume>
          (
          <year>2017</year>
          )
          <fpage>44</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>V. R.</given-names>
            <surname>Basili</surname>
          </string-name>
          , G. Caldiera,
          <string-name>
            <given-names>H. D.</given-names>
            <surname>Rombach</surname>
          </string-name>
          ,
          <article-title>The goal question metric approach</article-title>
          , in: J. J.
          <string-name>
            <surname>Marciniak</surname>
          </string-name>
          (Ed.),
          <source>Encyclopedia of Software Engineering</source>
          , John Wiley &amp; Sons, NY, USA,
          <year>1994</year>
          , pp.
          <fpage>528</fpage>
          -
          <lpage>532</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <article-title>Development of nasa-tlx (task load index): Results of empirical and theoretical research</article-title>
          , in: P. A.
          <string-name>
            <surname>Hancock</surname>
          </string-name>
          , N. Meshkati (Eds.),
          <source>Human Mental Workload</source>
          , volume
          <volume>52</volume>
          of Adv. Psychol., North-Holland,
          <year>1988</year>
          , pp.
          <fpage>139</fpage>
          -
          <lpage>183</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>F. S.</given-names>
            <surname>Lopez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Condori-Fernández</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Catalá</surname>
          </string-name>
          ,
          <article-title>Towards real-time automatic stress detection for ofice workplaces</article-title>
          , in: J. A.
          <string-name>
            <surname>Lossio-Ventura</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Muñante</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Alatrista-Salas</surname>
          </string-name>
          (Eds.),
          <source>Information Management and Big Data</source>
          , volume
          <volume>898</volume>
          of Commun.
          <source>Comput. Inf. Sci.</source>
          , Springer,
          <year>2018</year>
          , pp.
          <fpage>273</fpage>
          -
          <lpage>288</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>