<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Pupillary Response: Removing Screen Luminosity Effects for Clearer Implicit Feedback</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>CCS Concepts</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Figure 1. Evaluation experiments at</institution>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Tomas Juhaniak</institution>
          ,
          <addr-line>Patrik Hlavac, Robert Moro, Jakub Simko</addr-line>
          ,
          <institution>Maria Bielikova Faculty of Informatics and Information Technology, Slovak University of Technology in Bratislava</institution>
          ,
          <country country="SK">Slovakia</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Human-centered computing~User models • Human-centered computing~Web-based interaction • Information systems~ Personalization</institution>
        </aff>
      </contrib-group>
      <abstract>
        <p>Pupillary dilation measured by eye-tracking can be useful source of implicit feedback for system adaptation and personalization. For example, cognitive load or emotional excitation can be inferred from it. However, practical exploitation of this phenomenon (e.g., in adaptive systems or user studies) has been limited due to other factors that influence pupillary dilation, namely changing luminosity of device screen. In this work, we present a personalized pupillary dilation model, which is able to predict the effects of screen luminosity on participant's pupil diameter. This information is useful for tracking true effects of cognitive load or emotional excitation of users. We demonstrate our model in a controlled eyetracking study with 73 participants. Pupillary response; personalized model; luminosity; cognitive load.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. INTRODUCTION</title>
      <p>
        The collection of implicit feedback is a necessary precursor to user
modelling, adaptation, and personalization. Implicit feedback
streams are, however, influenced by multiple factors that obfuscate
the information we seek. One such case is eye-tracked pupillary
response (dilation) measurement. It is often done to measure user’s
cognitive load [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] and emotional excitation [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], both of which are
useful in adaptation and personalization scenarios.
      </p>
      <p>
        Unfortunately, pupil diameter is also influenced by other important
factor: luminosity of environment surrounding the user, especially
the stimulus that user is perceiving on the screen. Untracked
changes in environment luminosity may completely disable any
tracking of cognitive load or emotional excitation. In practical
scenarios, stimulus luminosity cannot be kept stable and changes
with every new screen. When the screen content is heterochromous,
even the changes in gaze fixations may trigger pupillary dilations.
Only few solutions exist to this problem. Simple, but impractical,
is to restrict the task design. By keeping the stimuli roughly on the
same luminosity level throughout the experiment, one can attribute
any pupillary response to cognitive load or emotional excitation [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ].
A more sophisticated method of Xu et al. [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] measures mean
pupillary diameters for time segments with constant stimuli
luminosity and bases the cognitive load detection on dilation
deviation. Another method tries to distinguish between abrupt
(mental-state-caused) and moderate (luminosity-caused) pupillary
responses [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. These methods allow more variety in stimulus
luminosity, but are still hindered when used for complex scenarios.
If we want to successfully filter out luminosity effects on pupillary
dilation, we need to predict them for arbitrary, heterogeneous
stimuli, accounting also for exact fixation points.
      </p>
    </sec>
    <sec id="sec-2">
      <title>2. PERSONALIZED MODEL OF</title>
    </sec>
    <sec id="sec-3">
      <title>PUPILLARY RESPONSE</title>
      <p>The contribution of this work is the personalized model of pupillary
response (PMPR). Given the screen bitmap, the model is able to
predict pupil diameter of the user in non-excited and non-loaded
mental state. The model also takes into account the position of
user’s gaze fixations (e.g., when user focus to darker area, the
predicted diameter increases). The model expects that the
environment luminosity is invariant.</p>
      <p>
        PMPR needs to be calibrated to each user, as diameter range of the
pupil is a personal physiological trait. During calibration, user
focuses on the center of the screen while he is presented with a
sequence of defined stimuli, each with different color, visual
structure and luminosity values (see figure 1). After the calibration,
the user can work with any stimulus required and the model will
predict pupillary responses caused by the stimuli luminosity
changes, which can be then used to correct the overall measured
pupillary response.
PMPR is based on two main concepts: (i) pupil reference curve and
(ii) model of fixation-based image luminosity. The pupil reference
curve models the user-specific behavior of the pupil with the
changing luminosity of the stimuli. It is based on the work of Ellis
[
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] and assumes that the pupil diameter linearly increases with the
decreasing luminosity of the stimulus. Therefore, it can be
modelled as:
  =  × (1 − 
) + 
(2)
where   is the pupil diameter,  is the stimulus luminosity from
the interval 〈0,1〉,  is the value of pupil diameter for the white
stimulus ( = 1), and  is the slope of the reference curve.
The idea behind the model of fixation-based image luminosity is
based on the anatomical distribution of the rods and cones in the
eye [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], which makes our vision sharpest at the foveal area (2° to
3° of visual angle) with the visual acuity steeply decreasing the
farther from the center towards parafoveal and peripheral area (by
90% at 40° of visual angle) [7]. This means that the perceived
luminosity of the image is influenced the most by the luminosity of
the part of the image that is in the center of the visual attention, i.e.,
it is fixated by the user. The further the area of the image is from
the fixation point, the lower is its addition to the overall luminosity
perceived by the user. Therefore, when computing perceived
luminosity of the stimulus bitmap, we modify each pixel’s
luminosity with 2D Gaussian kernel centered at fixation point.
Thus, we can formally define the personalized model of pupillary
response as the following triple:

= ( ,  ,  )
(6)
where and a and b are the parameters from the reference curve and
σ is a parameter of the 2D Gaussian kernel. During the calibration,
these parameters are trained on various abstract colored and shaped
stimuli (e.g. planes, circles, diagonal splits).
      </p>
      <p>The first part of calibration serves to calibrate the reference curve
(parameters a and b). All of the desired luminosity levels are
projected with the help of the plain monochromatic stimuli, in
several iterations with brightening and blanking phases through the
spectrum of gray. In second part of the calibration procedure, the σ
parameter of the 2D Gaussian kernel is trained. We project to the
users white circles on a black background with different diameters
(ranging from 1016px to 16px), while the calibrated person is
looking at their center (see Figure 2) and record the actual pupil
diameters. The numeric optimization method (quasi-newton BFGS
algorithm) is then used to find the optimal value of σ. The whole
calibration procedure takes about four minutes, which makes it
feasible to use before a user study that uses cognitive load or
emotional excitation as an indicator of implicit feedback.</p>
    </sec>
    <sec id="sec-4">
      <title>3. EVALUATION</title>
      <p>We have done a preliminary evaluation of our model. We invited
73 participants to participate in approximately four minute
eyetracking experiment. Each participant looked at a series of stimuli
which were expected to yield changes in the pupil diameter due to
changing luminosity. The stimuli series consisted of a set of plain
color images, two-color images and real web pages. During the
exposure, participants were asked to focus their sight on a cross in
the middle of the screen. We focused on the evaluation of the
precision of the pupil diameter predictions of the model.
The experiment was carried out in the UX Group laboratory (see
figure 1) of the User Experience and Interactions Research Center1
at our university, which contains 20 working stations each equipped
with Tobii X2-60 eye-tracker with 60Hz sampling frequency.
For each participant, we have trained their PMPR model (over
portion of plain color and two-color stimuli recordings) and tested
the prediction on the rest (rest of plain color images and web pages).
For both plain color stimuli and web pages, the relative prediction
error of our model was up to 5% (of the total dilation range of
participants). The highest error for plain color stimuli reached 10%,
for web pages 15%. Furthermore, the success rate of prediction
over web stimuli varied (for some participants, the model
consistently predicted better than for others).</p>
      <p>Importantly, the trained parameters of the model varied among the
participants, which justifies the whole concept of the personalized
model of pupillary response and our novel approach of the
individually trained reference curve.</p>
    </sec>
    <sec id="sec-5">
      <title>4. CONCLUSION</title>
      <p>The contribution of this work is a personalized model of pupillary
response (PMPR), which can predict changes of pupil diameter
caused by screen stimulus luminosity. The relative errors of
prediction reach about 5% of the total dilation range. We are
confident that this makes the model usable for successfully
separating effects of luminosity from overall pupillary dilation
effects and thus enables to measure implicit feedback such as
cognitive load or emotional excitation. Furthermore, PMPR works
over any type of stimuli including real web pages, and does not
need carefully crafted stimuli with constant luminosity. This eases
the design of studies, which aim to measure cognitive load or
emotional excitation using pupillary dilation tracking.
Acknowledgement. This work was partially supported by the
Scientific Grant Agency of the Slovak Republic, grant No. VG
1/0646/15, the Slovak Research and Development Agency under
the contract No. APVV-15-0508 and was created with the support
of the Ministry of Education, Science, Research and Sport of the
Slovak Republic within the Research and Development
Operational Programme for the project ITMS 26240220084,
cofunded by the European Regional Development Fund. The authors
would like to thank for financial assistance from the STU Grant
scheme for Support of Excellent Teams of Young Researchers.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Epps</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          :
          <article-title>A comparison of four methods for cognitive load measurement</article-title>
          .
          <source>In: Proc. of the 23rd Australian Computer-Human Interaction Conf., ACM</source>
          ,
          <year>2011</year>
          , pp.
          <fpage>76</fpage>
          -
          <lpage>79</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <surname>Ellis</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>The pupillary light reflex in normal subjects</article-title>
          .
          <source>British J. of Ophthalmology</source>
          ,
          <year>1981</year>
          , vol.
          <volume>65</volume>
          , no.
          <issue>11</issue>
          , pp.
          <fpage>754</fpage>
          -
          <lpage>759</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <surname>Marshall</surname>
            ,
            <given-names>S.P.:</given-names>
          </string-name>
          <article-title>The index of cognitive activity: Measuring cognitive workload</article-title>
          .
          <source>In: Human factors and power plants</source>
          ,
          <source>2002. Proc. of the 2002 IEEE 7th conf., EEE</source>
          ,
          <year>2002</year>
          , pp.
          <fpage>7</fpage>
          -
          <lpage>5</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <surname>Xu</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Choi</surname>
          </string-name>
          , E.:
          <article-title>Pupillary response based cognitive workload measurement under luminance changes</article-title>
          . In:
          <string-name>
            <surname>Human-Computer</surname>
            <given-names>Interaction-INTERACT</given-names>
          </string-name>
          <year>2011</year>
          . Springer,
          <year>2011</year>
          , pp.
          <fpage>178</fpage>
          -
          <lpage>185</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <surname>Partala</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          &amp;
          <string-name>
            <surname>Surakka</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <year>2003</year>
          .
          <article-title>Pupil size variation as an indication of affective processing</article-title>
          .
          <source>International Journal of Human Computer Studies</source>
          ,
          <volume>59</volume>
          (
          <issue>1-2</issue>
          ), pp.
          <fpage>185</fpage>
          -
          <lpage>198</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <surname>Irwin</surname>
            ,
            <given-names>D.E.</given-names>
          </string-name>
          ,
          <year>2004</year>
          .
          <article-title>Fixation Location and Fixation Duration as Indices of Cognitive Processing. The interface of language, vision, and action: Eye movements and the visual world</article-title>
          , (
          <volume>217</volume>
          ), pp.
          <fpage>105</fpage>
          -
          <lpage>133</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>