<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Re-testing the Perception of Social Annotations in Web Search</article-title>
      </title-group>
      <contrib-group>
        <aff id="aff0">
          <label>0</label>
          <institution>Ed H. Chi Google 1600 Amphitheatre Pkwy Mountain View</institution>
          ,
          <addr-line>CA 94043</addr-line>
          <country country="US">USA</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Jennifer Fernquist Google 1600 Amphitheatre Pkwy Mountain View</institution>
          ,
          <addr-line>CA 94043</addr-line>
          <country country="US">USA</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>We evaluated the perception of social annotations designed via guidelines recommended by Muralidharan, Gyongyi, Chi, 2012. The initial study found participants noticed the annotation only 11% of the time with annotations shown below the search result snippet. Our refined study revealed that the proposed design with the annotation above the snippet increased noticeability to 60%. Replication studies are often iterative version of old studies, and this was no exception. The new study refined the protocol for measuring 'notice' events, and modified the tasks to include tasks that are more relevant to recent news articles.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>Presented at RepliCHI2013. Copyright © 2013 for the individual papers
by the papers’ authors. Copying permitted only for private and academic
purposes. This volume is published and copyrighted by its editors.</p>
    </sec>
    <sec id="sec-2">
      <title>Author Keywords</title>
      <p>Annotation; social search; eyetracking; user study.</p>
    </sec>
    <sec id="sec-3">
      <title>ACM Classification Keywords</title>
      <p>H.5.m. Information interfaces and presentation (e.g.,
HCI): Miscellaneous</p>
    </sec>
    <sec id="sec-4">
      <title>Introduction</title>
      <p>
        The abundance of information on the web suggests the
importance of creating an environment in which users
have the appropriate signals to make decisions about
which search results are the most useful to them. As
more of the web involves social interactions, they
produce a wealth of signals for searching the most
interesting and relevant information. Much research has
been done on modifying search ranking based on social
signals for web pages [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ][
        <xref ref-type="bibr" rid="ref2">2</xref>
        ][
        <xref ref-type="bibr" rid="ref3">3</xref>
        ][
        <xref ref-type="bibr" rid="ref5">5</xref>
        ][
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], but how should
we present the social signals for web search results?
The most recent paper that we have found is the
CHI2012 paper on social annotations by Muralidharan
et al. [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ].
      </p>
    </sec>
    <sec id="sec-5">
      <title>Previous Research</title>
      <p>
        Muralidharan et al. [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] studied the perception of social
annotations appearing below search results, as in Figure
1. Consistent with prior papers, we use the term “social
signals” to refer to any social information that is used
to affect ranking, recommendation or presentation to
the user. We use the term “social annotations” to refer
to the presentation of social signals for an explanation
as to why a search or recommendation result is
presented. Thus, a social signal only becomes an
annotation when it is presented to the user.
      </p>
      <sec id="sec-5-1">
        <title>Study Protocol</title>
        <p>Their first study had two parts: (1) In the first part,
participants conducted 18-20 search tasks, randomly
ordered. Half were designed so that one or two social
annotations would appear in the top four or five results.
The search results pages were presented as static mocks
that were generated before the study, customized for
each participant.
(2) The second part consisted of a retrospective
thinkaloud (RTA) where they walked the participant through
each task using the eyetrace data post hoc. During the
interview, researchers checked noticeability by asking if
the participants noticed the social annotations, either by
them mentioning they saw them or being explicitly asked
if they had seen them. During the RTA the researchers
also obtained qualitative feedback about social
annotations.</p>
        <p>The second study compared the perception of multiple
designs of social annotations. They varied profile image
size (small, large), snippet length (1, 2, 4 lines), and
annotation position (above, below snippet). For this study
the same mocks were used for each participant, with
customization only for customizing familiar names and
faces of people in the annotations. In the second study,
noticeability of the annotations was measured by counting
the number of fixations.</p>
      </sec>
      <sec id="sec-5-2">
        <title>Findings</title>
        <p>In the first study, they found that only 5 of the 45 (11%)
of the visible social annotations were noticed. In the
second study, they found that there were fewer fixations
on annotations when: the snippet length was longer; the
image was smaller; and the annotation was below the
snippet. They concluded that the optimal design for a
social annotation is one with a large picture, above the
snippet, with a short snippet length.</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>Our Method and Replication</title>
      <p>We aimed to actually test the proposed annotation
design guidelines from study 2 using live user data to
see if people notice the annotations more by using the
method from study 1. Specifically, we wanted to test
with live data that is relevant to participants (from their
connections), as opposed to the static images used
previously. An example of a social annotation with the
new design is shown in Figure 2.</p>
      <sec id="sec-6-1">
        <title>Study Protocol</title>
        <p>Experimental sessions consisted of 3 parts, the first two
using essentially the same protocol as experiment 1’s in
the previous work, with some improvements.</p>
        <sec id="sec-6-1-1">
          <title>PART 1: SEARCH TASKS</title>
          <p>We designed planned 16-20 custom search tasks for
each subject, at least eight of which were “social
search" tasks designed to organically pull up social
annotations. The 8 non-social search tasks were the
same as used in the prior work.</p>
          <p>In order to ensure that personal results appear for as
many queries as required, we designed 2-4 additional
social search tasks for each participant that were
intended to bring up personal results. This way, if one
social search task did not bring up personal results, we
gave them the additional tasks to help ensure that they
saw 8 tasks with personal results.</p>
        </sec>
        <sec id="sec-6-1-2">
          <title>PART 2: RETROSPECTIVE THINK-ALOUD</title>
          <p>
            After the search tasks, we immediately conducted a
retrospective review of eye-tracking traces for search
tasks in which subjects exhibited behaviors of some
interest to the experimenter. Review of eye-tracking
videos prompted think-aloud question answering about
participants’ process on the entire task, particular
interesting pages, and particular interesting results.
Unlike Experiment 1 in Muralidharan et al. [
            <xref ref-type="bibr" rid="ref4">4</xref>
            ], we
examined the eyetrace data directly by hand to
determine noticeability, rather than through verbal
feedback during the RTA tasks.
          </p>
        </sec>
        <sec id="sec-6-1-3">
          <title>PART 3: THINK-ALOUD TASKS</title>
          <p>Finally, participants performed two or three different
search queries for which we determined ahead of time
that should bring up relevant personal results. Here we
gathered qualitative feedback on social annotations.</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-7">
      <title>Results</title>
      <p>In total, we collected eye-trace data for 153 tasks from
nine subjects. Each eye-trace data for each task was
analyzed by hand by an experimenter to understand:
which positions contained personal search results;
whether the search result was in the field of view in the
browser; and importantly, whether the subject fixated
on the result and/or the social annotation. This funnel
analysis approach is different than the previous work’s
approach of asking participants if they noticed the
annotations.</p>
      <p>
        We discovered that participants fixated on annotations
in 35 of the 58 tasks where they appeared (60%). This
is a dramatic improvement over the 11% perception
rate of the Muralidharan et al. [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. We account this
difference primarily to the new annotation design.
      </p>
    </sec>
    <sec id="sec-8">
      <title>Replication Discussion</title>
      <sec id="sec-8-1">
        <title>Access to Previous Experimental Data. We were able to</title>
        <p>repeat the exact same tasks performed in the previous
work but only because we share a co-author who had
access to the data. If anyone else tried to replicate the
study, they would not have been able to do so as
effectively.</p>
        <p>
          Temporal Challenges. Even though the search tasks
were identical, because the study was conducted
several months later, some of the task questions were
no longer topically relevant. For example, one task
asked “What is the website for the Google image
labeling game?” At the time of our study, the website
was no longer active. Similarly, the search task “Find
some information about the Nevada law legalizing
selfdriving cars” brought up news articles from the
previous summer, when Muralidharan et al. [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ]
conducted their research, since it was no longer recent
news.
        </p>
        <p>This raises a big issue for research replication:
changing environments such as time or space. In our
case, the tasks lost their relevancy over time.
Researchers could help mitigate this by rewriting tasks
so they are more relevant but still in the same vein as
the original. For example, we could have written a
different task that was more topical but would still be
categorized as news. It must be decided which would
cause the least amount of discrepancy for replication:
maintaining the identical, less relevant task or rewriting
a relevant task that differs from the original.</p>
        <p>Iteration and Refinement. The primary difference in our
protocol, measuring perception with fixation data rather
than verbal confirmation, offered an improvement to
the previous work.</p>
        <p>Even with those challenges, we feel that we were
successful in our replication efforts. We conducted an
almost identical study to confirm the proposed
improved design for social annotations and found a
large increase in perception.</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>Bao</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Xue</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wu</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yu</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fei</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Su</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          <article-title>Optimizing web search using social annotations</article-title>
          .
          <source>In Proc. WWW</source>
          <year>2007</year>
          ,
          <volume>501</volume>
          -
          <fpage>510</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <surname>Carmel</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zwerdling</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Guy</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ofek-Koifman</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          , Har'el, N.,
          <string-name>
            <surname>Ronen</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Uziel</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yogev</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Chernov</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <article-title>Personalized social search based on the user's social network</article-title>
          .
          <source>In Proc. CIKM</source>
          <year>2009</year>
          ,
          <volume>1227</volume>
          -
          <fpage>1236</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <surname>Heymann</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Koutrika</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Garcia-Molina</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <article-title>Can social bookmarking improve web search?</article-title>
          <source>In Proc. WSDM</source>
          <year>2008</year>
          ,
          <volume>195</volume>
          -
          <fpage>206</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <surname>Muralidharan</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gyongyi</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Chi</surname>
            ,
            <given-names>E. H.</given-names>
          </string-name>
          <article-title>Social annotations in web search</article-title>
          .
          <source>In Proc. CHI</source>
          <year>2012</year>
          ,
          <volume>1085</volume>
          -
          <fpage>1094</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <surname>Yanbe</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jatowt</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nakamura</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Tanaka</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <article-title>Can social bookmarking enhance search in the web?</article-title>
          <source>In Proc. JCDL</source>
          <year>2007</year>
          ,
          <volume>107</volume>
          -
          <fpage>116</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <surname>Zanardi</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Capra</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <article-title>Social ranking: uncovering relevant content using tag-based recommender systems</article-title>
          .
          <source>In Proc. RecSys</source>
          <year>2008</year>
          ,
          <fpage>51</fpage>
          -
          <lpage>58</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>