<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Faceted Search for Library Catalogs: Developing Grounded Tasks and Analyzing Eye-Tracking Data</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Robert Capra</string-name>
          <email>rcapra3@unc.edu</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Bill Kules</string-name>
          <email>kules@cua.edu</email>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Matt Banta</string-name>
          <email>matt.banta@gmail.com</email>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Tito Sierra</string-name>
          <email>tito_sierra@ncsu.edu</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>NCSU Libraries North Carolina State University Raleigh</institution>
          ,
          <addr-line>NC</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>School of Information and Library Science University of North Carolina Chapel Hill</institution>
          ,
          <addr-line>NC</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>School of Library and Information Science The Catholic University of America Washington</institution>
          ,
          <addr-line>DC</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2009</year>
      </pub-date>
      <fpage>19</fpage>
      <lpage>23</lpage>
      <abstract>
        <p>In this paper, we describe two aspects of a study we conducted of faceted search in an online public access library catalog (OPAC). First, we describe how we used log data from a university OPAC to develop a set of grounded tasks. Then, we describe our use of eye-tracking in a controlled laboratory setting to examine user behaviors performing the grounded tasks. We discuss the challenges we encountered both in using the log data to develop tasks and in collecting and analyzing the eye-tracking data.</p>
      </abstract>
      <kwd-group>
        <kwd>General Terms Measurement</kwd>
        <kwd>Experimentation</kwd>
        <kwd>Human Factors</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. INTRODUCTION</title>
      <p>
        Many libraries have recently redesigned their online public access
catalogs (OPACs) to include faceted metadata as part of the
search interface. In these systems, metadata such as the Library of
Congress subject headings, time period, and region are displayed
as facets that can be used to explore and refine search results (see
Figure 1). There are many open research questions about how
people use facets in a search process and the library science
community is especially interested in how these redesigned
OPACs are being used. We designed a study to examine how
long and in what sequences searchers looked at the major
elements of a faceted OPAC interface [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ].
      </p>
      <p>This paper describes two types of challenges encountered along
the way: developing exploratory search tasks and analyzing eye
tracking data.</p>
    </sec>
    <sec id="sec-2">
      <title>2. LOG ANALYSIS OF SEARCHES</title>
      <p>
        Our study needed search tasks that balanced two competing
needs: first, the tasks needed to induce an exploratory mode of
search instead of the directed mode used in many studies. Second,
the tasks needed to be constructed in a way that allowed us to
make comparisons between subjects. In addition, the tasks needed
to be appropriate for the catalog available on the test system,
which was based on the North Carolina State University (NCSU)
Libraries OPAC, reflecting real usage of that catalog. The online
library catalog for NCSU serves on average 7,824 search
transactions and 1,087 user sessions per day [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ].
      </p>
      <p>To develop the tasks, we extracted three days of anonymized log
data from the servers. This extracted data included both keyword
search terms and any facets used in the searches. Our goal was to
use this log data to identify actual searches executed on the NCSU
OPAC that made use of facets. We were especially interested in
identifying exploratory searches (as opposed to directed or
known-item searches) in the log data.</p>
      <p>We manually looked through the extracted log data to identify
situations in which the user appeared to be doing an exploratory
search that also included the use of facets. We looked for log
entries where it was clear that the user issued several searches
with the same or related keywords and in which they interacted
with the results. Our selection criteria required that the log file
show that the searcher: 1) had looked through more than one
page of results, 2) had selected more than one facet that was not
identical to the search term, and 3) the selected facets were from
the subject, time period, and region facets. The deployed NCSU
OPAC has additional facets, but we decided to focus on only
these three for our study.</p>
      <p>
        To further define the tasks, we then conducted our own searches
using the topics that were extracted from the log files. If a single
keyword search could easily address the topic, it was either
rejected as too easy or modified to either broaden or narrow its
scope. Iterating this process, we developed a set of four
exploratory search tasks to use in the study. More details of our
task development and refinement process are given in [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ].
There are obvious difficulties in isolating exploratory searches
from log data. First, the log data did not link queries across
sessions, so there was no way of knowing with 100% certainty
that two queries were done by the same user. However, we often
observed sequences of closely related search terms in close time
proximity that indicated an exploratory style search. Second, it is
often impossible to know the exact motivations behind the actions
observed in the log data. For example, what was the underlying
task that lead a searcher to issue the query? Why did they chose
to click on that facet? However, for our purposes, the log data
provided a rich set of indicators to use in developing a set of
exploratory search tasks grounded in real-world searches.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. COLLECTING EYE-TRACKING DATA</title>
    </sec>
    <sec id="sec-4">
      <title>3.1 Interface Design</title>
      <p>We used a Tobii 2150 remote eye tracker (http://www.tobii.com)
to collect the eye-tracking data. This system includes a 21” LCD
monitor with embedded infrared cameras that sample at 50Hz.
The monitor resolution was set to 1024x768.</p>
      <p>For this study, we focused on how facets were used in the search
process. The deployed NCSU OPAC includes many interface
elements and features that are tangential to our current research
interest. To keep the study focused, we developed a customized
OPAC interface, shown in Figure 1. There were six major areas
of interest (AOIs) in the interface: 1) a keyword query search
box, 2) an area to display the facets, 3) a breadcrumb trail
showing the current search terms and selected facets, 4) an area to
display the results list, 5) a drop-down menu to select how to sort
the results, and 6) a checkbox for each result so that the user
could indicate which results they wanted to record as their
“answers” for each task. This customized interface still accessed
the full NCSU catalog of over 1.8 million records.</p>
      <p>To facilitate collecting the eye-tracking data, we made several
adjustments to this customized interface. First, we made sure that
the interface used fixed-width elements when possible so that we
could easily define a template for the areas of interest on each
page. Second, we included 5 pixels of “padding” between
interface elements to help increase the precision of gaze data
collection for specific AOIs.</p>
    </sec>
    <sec id="sec-5">
      <title>3.2 Data Collection</title>
      <p>Data collection using the eye-tracker was a tricky process. First,
we seated each participant at the computer with the eye-tracker
and went through a calibration process. After the first and second
tasks, while the participant was completing a post-task
questionnaire, the experimenter would quickly skim a video that
showed the eye-traces that were captured from the previous task
to make sure that the eye-tracking was good. In cases where it
had problems, we would either recalibrate the equipment and/or
remind the participant to sit as they had been sitting when doing
the calibration. For two participants, the equipment could not
maintain tracking for more than a few seconds and we had to
discard the tracking data.</p>
      <p>We observed that changes of posture were often the cause of
eyetracking failure. A typical example was that participants would sit
in a neutral posture while doing the calibration, but then either
slump or “lean in” while engaged in the tasks. We often had to
gently remind participants during the tasks to resume their
original posture. While we initially were reluctant to interrupt
them to correct their posture, we believe that the negative impact
of this interruption was very small compared to the gains in better
eye-tracking. We often used wording to encourage the participant
to help us, such as, “The equipment is being finicky today, could
you just sit up a bit so it can track you better?” Other types of
eye-tracking such as head-mounted units might not have these
issues with posture causing a loss of tracking.</p>
      <p>The challenge of maintaining tracking has encouraged us to
consider using a secondary monitor that will display the tracking
status in our subsequent studies. This will allow us to monitor the
tracking in real-time during the tasks and to encourage the
participants to adjust their posture if needed.</p>
      <p>One other challenge encountered was caused by the automatic
update feature of Microsoft Windows. During the course of data
collection (which spanned a week), the system performed an
automatic update which upgraded the Internet Explorer browser to
version 7. This was not compatible with the Tobii eye tracker and
forced us to reschedule several sessions while we downgraded
back to IE6.</p>
    </sec>
    <sec id="sec-6">
      <title>4. ANALYZING EYE-TRACK DATA</title>
      <p>
        Tobii Clearview analysis software (v 2.7.1) was used to segment
each web page viewed into the areas of interest (AOIs). This was
a labor intensive step. Each web page viewed had to be
segmented by hand by defining a box around each AOI using a
GUI tool. Templates can be used to define the locations of fixed
size and fixed position AOIs. However, for each page, the AOIs
from this template had to be adjusted because some of the
interface elements were of variable size (both horizontal and
vertical). For example, the vertical size of the facet AOI
depended on the number and length of the facets. Cutrell et al.
[
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] overcame a similar problem by embedding custom JavaScript
code in the web pages they were studying that automatically
extracted the locations and dimensions of bounding boxes based
on the Document Object Model (DOM) of the page. These
dimensions could then be used to automatically generate the AOIs
definitions.
      </p>
      <p>We analyzed the raw eye-gaze data to extract fixations that had a
minimum of 100ms duration within a radius of 30 pixels.
Different domains use different fixation criteria. For example, for
reading text, fixations may be more tightly defined than for
imageoriented tasks such as visual search. For reading tasks, the
manufacturer (Tobii) recommended a 20 pixel radius for 40ms.
For image tasks, they recommend a 50 pixel radius for 200ms.
Because our tasks involved both aspects of reading and visual
search, we chose their recommendations for mixed content (30
pixel radius for 100ms duration).</p>
      <p>After defining the AOIs and extracting the fixations, the Tobii
software output a time ordered sequence of gaze data. We wrote
scripts in PHP to convert and analyze this data. The scripts had to
accumulate fixations across AOIs, tasks and individual page
views.</p>
      <p>
        In analyzing eye-tracking data, two measures have been widely
used for related studies: fixation counts and fixation times.
Fixation count is thought to be an indicator of the importance of
the item (or AOI) being fixated upon [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. Fixation time is
considered to be an indicator of the complexity of the element.
We initially focused on analysis of the cumulative fixation time
for each AOI, but became interested in the transitions between
AOIs to examine the pattern of eye movement on the page.
Specifically, we extracted “gaze transition pairs” between AOIs
for all participants, task scenarios, and page views. We used this
data to generate directed graphs to summarize the most commonly
occurring gaze paths between AOIs. An example graph is shown
in Figure 2. This technique allowed us to see that many
transitions occurred between the results and facets and between
the results and breadcrumb area. We believe that directed graph
summarization shows great promise as an eye-tracking data
analysis tool.
      </p>
    </sec>
    <sec id="sec-7">
      <title>5. ACKNOWLEDGMENTS</title>
      <p>This work was supported in part by grants from the NSF/Library
of Congress (ISS 0455970 and ISS 0812363) and a grant from the
Catholic University Grant-in-Aid Committee. We thank Doug
Oard for the use of the eye tracker, and Joseph Ryan and Jason
Casden for their help in configuring the interface for this study.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>Lown</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          (
          <year>2008</year>
          ).
          <article-title>A Transaction Log Analysis of NCSU's Facted Navigation OPAC</article-title>
          .
          <article-title>Master's paper</article-title>
          ,
          <source>School of Information and Library Science</source>
          , University of North Carolina at Chapel Hill,
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <surname>Kules</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Capra</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Banta</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sierra</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          (
          <year>2009</year>
          ).
          <article-title>What Do Exploratory Searchers Look at in a Faceted Search Interface? In Proceedings of the 9th ACM/IEEE-CS Joint Conference on Digital libraries (Austin</article-title>
          , TX, June 15 - 19,
          <year>2009</year>
          ).
          <article-title>JCDL 2009</article-title>
          . ACM Press, New York, NY.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <surname>Cutrell</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Guan</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          <year>2007</year>
          .
          <article-title>What are you looking for?: an eye-tracking study of information usage in web search</article-title>
          .
          <source>In Proceedings of theSIGCHI Conference on Human Factors in Computing Systems</source>
          (San Jose, California, USA, April 28 - May 03,
          <year>2007</year>
          ).
          <source>CHI '07. ACM</source>
          , New York, NY,
          <fpage>407</fpage>
          -
          <lpage>416</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <surname>Jacob</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          &amp; and
          <string-name>
            <surname>Karn</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          (
          <year>2003</year>
          )
          <article-title>Eye Tracking in HumanComputer Interaction and Usability Research: Ready to Deliver the Promises (Section Commentary)</article-title>
          ,
          <source>in The Mind's Eye: Cognitive and Applied Aspects of Eye Movement Research</source>
          , ed. by J.
          <string-name>
            <surname>Hyona</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Radach</surname>
            , and
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Deubel</surname>
          </string-name>
          , pp.
          <fpage>573</fpage>
          -
          <lpage>605</lpage>
          , Amsterdam, Elsevier Science.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>