<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Proceedings+of+the+CHI2013(Workshop(! on#the#Replication#of#HCI#Research!</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Position!Papers!</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>University-of-Nottingham,-UK- - David$Coyle$</institution>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2013</year>
      </pub-date>
      <abstract>
        <p>RepliCHI2013-was-the-first-workshop7style-event-in-the-developing-RepliCHI-agenda,-and-thus-the-first-to-accept-papers-aboutreplications-as-submissions-from-the-HCI-community.-Three-position-papers-were-accepted,-including-positions-that-were-both- for-and-against-the-view-that-replication-of-research-is-important-in-our-field;-in-particular-these-position-papers-helped-to-bothframe-and,-for-the-first-time,-explore-the-limits-of-the-agenda.-Ten-Experience-Reports-then-helped-discuss-the-very-nature-of- replication-in-the-HCI-community-through-grounded-examples-and-real-events.-To-support-discussion,-the-workshop-welcomedfive-“original-authors”-of-work-that-was-replicated-in-our-submissions,-which-enriched-the-insights-gained-particularly-when- replicating-other-peoples-work,-and-trying-to-be-replicable-when-publishing.-For-an-outcome,-the-workshop-focused-on-furtherrefining-a)-our-understanding-of-HCI-replication-practices,-and-b)-the-nature-of-the-RepliCHI-event-for-subsequent-years.-</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Executive!Summary!</title>
      <sec id="sec-1-1">
        <title>Max$L.$Wilson$</title>
      </sec>
    </sec>
    <sec id="sec-2">
      <title>Organised!by!</title>
      <p>Ed$H.$Chi$</p>
      <sec id="sec-2-1">
        <title>Paul$Resnick$</title>
        <p>Page$3$=$$ Avoid$“It’s$JUST$a$Replication”$</p>
        <p>Bonnie&amp;E.&amp;Joh&amp;n
Page$8$=$$ Is$replication$important$for$HCI?$</p>
        <p>Christian&amp;Greiffenhagen&amp;&amp;&amp;Stuart&amp;Reeve&amp;s
Page$14$=$$ RepliPRI:$Challenges$in$Replicating$Studies$of$Online$Privacy$</p>
        <p>Sameer&amp;Pati&amp;l
Experience!Reports!–!Presented!in!Tandem!with!Original!Authors!
Page$19$=$$ Replicating$an$International$Survey$on$User$Experience:$Challenges,$Successes$and$Limitations&amp;</p>
        <p>Carine&amp;Lallemand,&amp;Vincent&amp;Koenig&amp;G&amp;u&amp;illaume&amp;Gronie&amp;r
Page$24$=$$ Replicating$and$Extending$Research$on$Relations$between$Visual$Aesthetics$and$Usability&amp;</p>
        <p>Noam&amp;Tractinsky&amp;
Page$29$=$$ Replicating$and$Extending$a$Facebook$Uses$&amp;$Gratifications$Study:$Five$Years$Later&amp;</p>
        <p>Tasos&amp;Spiliotopoulos&amp;&amp;&amp;Ian&amp;Oakley&amp;
Page$34$=$$ NewsCube$Replication:$Experience$Report$</p>
        <p>Sidharth&amp;Chhabra&amp;&amp;&amp;Paul&amp;Resnick&amp;
Page$39$=$$ Teaching$HCI$Methods:$Replicating$a$Study$of$Collaborative$Search&amp;</p>
        <p>Max&amp;L.&amp;Wilso$ n
Experience!Reports!
Page$44$=$$ Do$lab$effects$transfer$into$the$real=world?$And$should$we$care?$</p>
        <p>Petr&amp;Slovak,&amp;Paul&amp;Tenne&amp;&amp;n&amp;tGeraldine&amp;Fitzpatrick&amp;
Page$49$=$$ Re=testing$the$Perception$of$Social$Annotations$in$Web$Search$</p>
        <p>Jennifer&amp;Fernquist&amp;&amp;&amp;Ed&amp;H.&amp;C&amp;hi
Page$53$=$$ Challenges$of$Replicating$Empirical$Studies$with$Children$in$HCI$</p>
        <p>Quincy&amp;Brown,&amp;Lisa&amp;Anthony,&amp;Robinew&amp;Berr,&amp;Germaine&amp;Irwin,&amp;Jaye&amp;Nia&amp;Bse&amp;r&amp;thel&amp;Tate&amp;
Page$58$=$$ Replicating$Residential$Sustainability$Study$in$Urban$India$</p>
        <p>Mohit&amp;Jain,&amp;Yedendra&amp;B.&amp;Shrinivasa&amp;Tna&amp;w&amp;anna&amp;Dillahun&amp;t
Page$63$=$ Replicating$and$Applying$a$Neuro=Cognitive$Experimental$Technique$in$HCI$Research$</p>
        <p>David&amp;Coyle&amp;
Page$67$=$$ Replicating$Two$TelePresence$Camera$Depth=of=Field$Settings$in$One$User$Experience$Study$
Jennifer&amp;Lee&amp;Carlson,&amp;Mike&amp;Paget&amp;&amp;&amp;Tim&amp;Mc&amp;Collum</p>
        <p>Avoiding “Itʼs JUST a Replication”
Bonnie E. John
IBM T. J. Watson Research Center
1101 Kitchawan Rd
Yorktown Heights, NY 10598 USA
bejohn@us.ibm.com
Presented at RepliCHI2013. Copyright © 2013 for the individual
papers by the papers’ authors. Copying permitted only for private and
academic purposes. This volume is published and copyrighted by its
editors.</p>
        <sec id="sec-2-1-1">
          <title>Abstract</title>
          <p>This position paper explores my experiences getting
replication studies accepted at the CHI conference over
the past 30 years. These experiences lead to my
hypothesis that CHI reviewers and program committee
members at all levels need education and technology
support to understand and appropriately consider
replication studies for publication at CHI. I propose a
draconian “zeroth iteration” on a design for extensions
to the Precision Conference System to spur discussion
about how we can design our values into our processes.</p>
        </sec>
        <sec id="sec-2-1-2">
          <title>Author Keywords</title>
          <p>Experimental design, replication.</p>
        </sec>
        <sec id="sec-2-1-3">
          <title>ACM Classification Keywords</title>
          <p>H5.m. Information interfaces and presentation (e.g.,
HCI): Miscellaneous.</p>
        </sec>
        <sec id="sec-2-1-4">
          <title>General Terms</title>
          <p>Human Factors</p>
        </sec>
        <sec id="sec-2-1-5">
          <title>Introduction</title>
          <p>Replication has been at the heart of science for as long
as the scientific method has existed; sometimes it feels
as though I have been fighting for the value of
replication at CHI almost as long. As an engineer by
training and inclination, replication is of even more
importance for the practice of UI design, in my view,
because practitioners can (and should) only trust
results from science when the results have been
replicated at several different research groups (i.e.,
direct replication) and the boundaries of applicability
have been thoroughly explored through
replicate+extend studies. I cannot count the number of
times I have heard “Reject; it’s JUST another Fitts’s
Law study” or Reject; it’s JUST another GOMS study” at
program committee meetings in our field. When
present, I have sometimes been able to rescue these
contributions to our field’s science base. I can only
imagine how many such papers were rejected when I,
or like-minded researchers, were not present and how
many potentially-contributing authors have been
discouraged by such “JUST a replication” reviews. This
position paper is a proposal of how to avoid “It’s JUST a
replication” in the absence of dogmatic senior
researchers like me.</p>
        </sec>
        <sec id="sec-2-1-6">
          <title>Hypotheses about the problem</title>
          <p>
            It is my experience that some sorts of replication are
more acceptable to reviewers and program committees
than others. The most acceptable seem to be those
that replicate only a method, e.g., Baskin and John [
            <xref ref-type="bibr" rid="ref1">1</xref>
            ]
used the same method of achieving extremely skilled
task execution performance as did Card, Moran and
Newell [
            <xref ref-type="bibr" rid="ref2 ref3">2</xref>
            ]. Using the same method to study
performance on a GUI CAD system [
            <xref ref-type="bibr" rid="ref1">1</xref>
            ] and a
command-line text editor [
            <xref ref-type="bibr" rid="ref2 ref3">2</xref>
            ] was not criticized by
reviewers, seemingly because the tasks were
sufficiently different. My hypothesis is that method
replication is not a problem in HCI research publication,
so much so that it might not even be recognized as a
type of replication.
          </p>
          <p>However, I know of replicate and extend papers falling
(or being pushed) into the JUST-a-replication barrel
when they vary any one of the myriad other variables
in a study.</p>
          <p>Extending the participants to a new user group.
For example, a study I cannot name for confidentiality
purposes was rejected when it replicated an educational
treatment using participants who were different from
the previously published work: they were at a
lesserknown school, they were in a different major and
therefore could be assumed to be less motivated to do
well on a topic, and were given less direct access to
expert support in doing the experimental support. The
fact that these participants performed as well as the
majors at a top-of-the-line school studying under the
inventor of the educational treatment is a replication
worth printing because it gives hope that the
educational treatment will scale beyond the reach of its
inventor.</p>
          <p>Similarly, a paper that was rescued from
JUST-areplication, but which I will not name to maintain
confidentiality, described a well-known HCI method
being used by practitioners far outside the HCI field,
having picked up the technique from the HCI literature
and made profitable use of it, verified with empirical
data. That any of our methods can be of use to people
without our help is a result worth publishing because it
also shows that the beneficial impact of our field can
extend beyond the reach of our limited number of
researchers.</p>
          <p>Extending the measures in the study to cover new
questions
Again, in a rejected paper I cannot reveal, a replication
was done that included additional survey data that
explored why some behavior was observed in both the
original and replication studies. The survey instrument
was new, the data was new, and, to me, the insight it
revealed was new, but this was rejected as
JUST-areplication. Thus, there seems to be a disagreement in
our community about how much extension constitutes
a publishable extension. In my opinion, the replication
itself was valuable and the extension was icing on the
cake, but that was not the opinion of the reviewers.
Differences of opinion about what does and does not
constitute a publishable contribution are not
uncommon, and in fact should be encouraged, but the
reviews did not even acknowledge that there was any
extension at all, causing me to hypothesize that the
definition of replicate+extend is not well assimilated
into our review community.</p>
          <p>
            Direct replication to increase statistical power so that
new questions can be answered
Tired of not being able to give details of the papers I
have discussed above, I offer my own rejected CHI
paper to make a point about direct replication [
            <xref ref-type="bibr" rid="ref5">4</xref>
            ]. We
had done a study with only six participants per
condition and the effect was so strong that it attained
statistical significance on some coarse measures and
was published at the IEEE’s International Conference on
Software Engineering [
            <xref ref-type="bibr" rid="ref4">3</xref>
            ]. The coarse measures did not
help us understand why the participants performed
better on some conditions than others and did not
distinguish between two conditions that had important
implications for the practical use of the technique we
were investigating. Therefore, we did a direct
replication of the previous study, justified combining
the data, and were able to tease out several new
insights given the increased power of the combined
study. We thought the results were a significant
contribution beyond the initial study, and in fact, these
results are the only ones that excite software
engineering audiences when I talk about them (SEs are
the target “users” of these research results).
          </p>
          <p>Whether you agree that the results are exciting enough
to publish is immaterial to the reviews we received –
“Reject; it’s JUST a replication” without comment on
the new analyses and results. This leads me to the
hypothesis that new analyses are not sufficiently valued
or understood by our reviewing community to warrant
comment. The replication “surface structure” is enough
to push a paper into the JUST-a-replication barrel.
And interesting point about the interaction of replication
and anonymous reviewing was brought out by this
paper as well. This was in the era of CHI’s strict rules
about anonymization, so we wrote about ourselves in
the third person, as instructed. A reviewer seemed to
think that using “Golden et. al’s” materials was
somehow cheating or lazy and criticized us for not
creating our own materials. Again, this leads to the
hypothesis that our reviewing community is in need of
education about the process of a good replication (i.e.,
NOT making your own materials) and highlights a
potential confound between anonyminity and
replication. Might the paper have been less harshly
reviewed if the reader had known that we did the
original study, i.e., we did do the hard work of creating
the materials and were not cheating or lazy?</p>
        </sec>
        <sec id="sec-2-1-7">
          <title>A proposed approach to a solution</title>
          <p>As explained above, my experiences lead me to the
hypothesis that if our community is to embrace
replication and publish good ones, reviewers need to be
educated about what makes a good replication and its
value to the field.</p>
          <p>It is not sufficient to instruct Associate Chairs (ACs) and
Sub-committee Chairs (SCs) as was done at the
Program Committee meeting for CHI2013, because
reviewer scores push replications down in the rankings
and we cannot depend on human memory in the heat
of PC debates to raise such papers to the level of
discussion.</p>
          <p>Therefore, I propose that we build our values into
submission and reviewing software (Precision
Conference System, PCS), to be a “job aid” to authors,
reviewers, ACs and SCs, delivering education at the
time it is needed. Below I present “iteration 0” of a
design for these extensions to PCS.</p>
          <p>Job aid for authors:
Present a required radio button for authors at
submission time. Include an information button next to
the question that leads to information about what a
replication study is and what the criteria for reviewing
are for a replication study.</p>
          <p>It is possible that we would want to ask for the type of
replication (direct replication, replicate+extend, or
conceptual replication), but that may be introducing too
much complexity in the first iteration.</p>
          <p>Job aid for reviewers
If the author has declared the paper to be a replication
study, then the review form shown to reviewers
changes to include specific required fields that apply to
replication studies. Include an information button next
to every field so the reviewer can get information about
acceptable replication processes and the general value
of replication at the time of filling out the review.
Depending on how much we believe our target users
need the education, we may consider presenting this
information in a modal dialog box when field is first
clicked by a reviewer with a button that dismisses the
dialog box and a checkbox “do not show me this again”
appearing after a reasonable amount of time needed to
read the text in the box.</p>
          <p>Reviewers should be able to identify themselves to PCS
as being skilled in assessing replications and interested
in doing so.</p>
          <p>Job aid for Associate Chairs (ACs)
If the author has declared the paper to be a replication,
this is indicated to the AC at paper-assignment time, so
the AC is aware that reviewers skilled in experimental
design and analysis should be recruited. Such
reviewers may be self-identified in PCS, as above. We
may also consider allowing ACs and SCs to identify
especially skilled replication reviewers in PCS, like we
currently acknowledge excellent reviews.</p>
          <p>At review time, the AC’s meta-review form also
changes to include required fields that specifically
address issues with replication, with information
buttons.</p>
          <p>PCS could also automatically mark this paper “to be
discussed at the PC meeting”. Depending on how
aggressive the CHI conference wants to be that year for
considering replication papers, this status may or may
not be changed by the AC.</p>
          <p>Job aid for Subcommittee Chairs (SCs)
If the author has declared the paper to be a replication,
this is indicated to the SC at the time that papers are
assigned to ACs, so the SC can assign an AC skilled in
assessing replication. When recruiting ACs for a
subcommittee likely to get replication submissions, the
SCs might be asked to identify one or two ACs who are
skilled in assessing replications, which will get the SCs
thinking about this necessary skill when they can do
something about it instead of when replication studies
arrive.</p>
          <p>At the PC meeting, the SC’s view should highlight the
papers that were identified by their authors as being
replication studies, so the SC can query the AC about
them during the meeting. Even if PCS allows the AC to
change the status of the paper to “do not discuss” it
would contribute to the education of all ACs if a
sentence or two were said at the PC meeting about why
this replication paper was not being discussed.</p>
        </sec>
        <sec id="sec-2-1-8">
          <title>Conclusion</title>
          <p>The zeroth iteration on changes to PCS proposed above
are purposely draconian to start discussion of how our
conference reviewing technology can support our value
system surrounding replication studies. I believe the
need is there, let’s put our UI design skills and our
SIG’s money where our values are.</p>
        </sec>
        <sec id="sec-2-1-9">
          <title>Acknowledgements</title>
          <p>This research was supported by in part by IBM. The
views and conclusions in this paper are those of the
authors and should not be interpreted as representing
the official policies, either expressed or implied of IBM.</p>
          <p>Is replication important for HCI?
Christian Greiffenhagen
Loughborough University
c.greiffenhagen@lboro.ac.uk
Stuart Reeves
University of Nottingham
stuart@tropic.org.uk</p>
          <p>Abstract
Replication is emerging as a key concern within
subsections of the HCI community. In this paper, we
explore the relevance of science and technology studies
(STS), which has addressed replication in various ways.</p>
          <p>Informed by this literature, we examine HCI’s current
relationship to replication and provide a set of
recommendations and points of clarification that a
replication agenda in HCI should concern itself with.</p>
          <p>Author Keywords
Replication; psychology; science and technology
studies; philosophy of science.</p>
          <p>ACM Classification Keywords
H.5.m. Information interfaces and presentation (e.g.,
HCI): Miscellaneous.</p>
          <p>Introduction
Replication is emerging as a concern within subsections
of the HCI community. A key motivation for this is a
feeling that HCI emphasises novelty over consolidation
of research; consolidation that can be achieved via
replication. In response, we advocate the relevance to
HCI of understandings of ‘replication’ emerging from
the philosophy and sociology of science and technology.</p>
          <p>This paper highlights a collection of rejoinders to the
ways in which this programme for replication is
currently conceptualised within HCI. In doing so we
intend to help the development of an endogenous
understanding of replication as a practice that can be a)
motivated, b) mature and c) fit for the purposes of HCI.</p>
          <p>Replication: Lessons from STS
We believe that debate on replication in HCI can be
enriched by STS and philosophy and sociology of
science. In this section we review some of the findings
of this literature and their pertinence to HCI.</p>
          <p>One of the motivations for replication within HCI is the
wish to make HCI more scientific by modelling HCI on
other sciences (e.g., “psychology, physics and
medicine” [11]). While there is nothing problematic in
asking for a field to involve more replication, to frame
this in terms of making it more ‘scientific’ is possibly
based on a mythical view of ‘good science’ of which
"[r]eplication of research is a cornerstone" [11]. This
view suggests this ‘science’ may be a homogenous
practice, possibly even based around a particular
method, ‘the’ scientific method. It also tends to think
about replication more from the perspective of the
philosophy of science, rather than the practice of
different sciences.</p>
          <p>
            In contrast, philosophical and sociological studies have
shown that ‘science’ refers to a fragile structure of
multiple disciplines and multiple methods linked by
‘family resemblances’ only [
            <xref ref-type="bibr" rid="ref4">9, 3</xref>
            ]. Not all empirical
sciences work with experiments, and the role of
experiments may differ between different fields.
          </p>
          <p>
            Complicating this picture is the separation between these
varied and autochthonous scientific practices and their
rendering into literature. Scientific literature is written in
such a way to as to promise replicability, emerging from
Boyle’s attempts to create scientific records that were
publicly accountable and would let ‘anyone’ replicate
experimental practices [10]. However, the nature of
instructions is such that they are always incomplete [
            <xref ref-type="bibr" rid="ref5">4</xref>
            ],
thus scientific instructions must be ‘filled in’ by competent
members of the target scientific community in order to
enact them as replications. This is one of the reasons why
Medawar characterised the scientific paper, somewhat
misleadingly, as a ‘fraud’ [
            <xref ref-type="bibr" rid="ref9">8</xref>
            ].
          </p>
          <p>
            STS reports an alternative view on the nature of
replication in the natural sciences to the surface view of
scientific replication where scientific articles (in
particular: their ‘method sections’) provide an adequate
instruction manual for replication work. Specifically it
problematises the notion of a ‘decisive experiment’ or
by extension a ‘decisive replication’. At the heart of this
problem is what Collins calls the “experimenter’s
regress” [
            <xref ref-type="bibr" rid="ref1">1</xref>
            ], that is, a circular relation between
experimental findings and the instruments used to
produce them. Reliable experimental findings
themselves rely upon reliable instruments and vice
versa. As a result, a key difficulty of replication and the
experimenter’s regress is that, particularly for
contested science, there is not necessarily any standard
for what is to be considered a valid replication. This
raises a principle problem, since it is not clear whether
a ‘failed’ replication is due to a problem with the
original experiment or the subsequent replication (“it is
often hard to tell whether an inability to replicate a
result is due to a group’s failings or a flaw in the
original paper” [5, p. 345]).
          </p>
          <p>
            Further to this, when we consider the track record of
replication in the natural sciences, STS literature argues
that replication in the (natural) sciences employs
replication for specific, highly motivated and reasoned
ends. Thus we find a marked absence of large amounts
of replication in the sciences unless we focus on particular
issues [1, pp. 210-211]. For instance, Collins’ tracing of
the construction of gravitational wave detectors during
the 1970s reveals the relevance of replication as an
activity for working through what was a contested,
controversial domain [
            <xref ref-type="bibr" rid="ref2 ref3">2</xref>
            ]. In short, ‘doing replication’ is
not always seen as a fundamental prerequisite for valid
scientific practice, since a vast number of results go
unreplicated: instead it emerges as the result of
pragmatic action for specific contested cases.
          </p>
          <p>In summary, then, our cursory examination of STS and
its related literature highlights that: a) there is no
singular form of science or scientific method upon
which to model; b) there is no ‘algorithmic’ method for
replicating directly from scientific literature (indeed,
this is not its purpose); c) ‘absolute’ security of results
is problematic in light of the experimenter’s regress;
and d) sciences often do not involve replication as a
‘matter of course’, it being difficult and of little value
unless motivated (typically via contestation of results).</p>
          <p>Replication within HCI
This issue of replication has become a centre of
discussion within HCI. In light of STS’s view on
replication, we seek to ask what is at stake in this
discussion. Why replicate? Or: What are the (different)
aims and motivations for replication?
Within HCI, it has been acknowledged that there is not
just one kind of replication. For example, Wilson et al.
distinguish between four forms: “direct replication”
(“driven by the aspirations of strong science”),
“conceptual replication” (replication via “alternative
methods”), “replicate &amp; extend” (building on prior
studies incrementally) and “applied case studies”
(replication through application of prior work) [11].</p>
          <p>Nuancing this view, we want to start with introducing
two different kinds of distinction to help us to think
about replication.</p>
          <p>The first distinction is between what we characterise as
textbook replication and frontier replication. By
‘textbook replication’ we refer to replications of
wellknown studies that are conducted from HCI textbooks,
typically as part of undergraduate or graduate
education. For instance, these could be replications of
well-known usability studies. We distinguish this from
‘frontier replication’ by which we mean replications of
‘ongoing’ or ‘recent’ studies. We see these forms as
conceptually and practically incommensurate, as
opposed to integral facets (e.g., see position in [11] on
“Benefits of Replication”). Thus, while the primary aim
and motivation of textbook replications is learning, the
point of frontier replication is often a form of ‘checking’
(which may even be done during the review process).</p>
          <p>As such we argue that the activities at this ‘frontier’
becomes the main issue for replication rather than what
is happening ‘in the textbook’.</p>
          <p>A second distinction has to do with what may be
replicable and what is actually replic-ated, in which the aims
for each are quite different. ‘Being replicated’ concerns
the ‘factual’ question of whether a particular study has,
actually, been replicated by other researchers or not.</p>
          <p>
            We say ‘factual’ since subsequent studies may or may
not be seen as valid replications, as in Collins’ study of
gravitational wave detectors [
            <xref ref-type="bibr" rid="ref2 ref3">2</xref>
            ]. We also note again
that a lack of actual replications may be related to
matters such as experiments being too costly, too time
consuming or lacking in providing the experimenter any
obvious credit.
          </p>
          <p>In contrast ‘being replicable’ is motivated by the ‘in
principle’ possibility of some other researcher being
able to replicate an empirical study. This is often cited
as one of the differences between ‘quantitative’ and
‘qualitative’ methods (very problematic descriptions
themselves), where the former supposedly produce
results that could be replicated (again, ‘in principle’),
while the later are not. For instance, ethnographic
research is often said to be too reliant on the
‘subjective’ insights of the ethnographer, resulting in
non-generic and non-replicable findings.</p>
          <p>
            What’s at stake in this distinction? We would argue that
the issue of ‘being replicable’ concerns a foundational
question, in particular, whether HCI is a science and its
preferences for particular methods over others. These
questions are not new: psychology—which has strongly
informed HCI’s development—has repeatedly
foregrounded replication as an explicit agenda, such as
in response to perceived experimental biases (e.g.,
being too ‘WEIRD’ [
            <xref ref-type="bibr" rid="ref7">6</xref>
            ]), as well as intentional and
unintentional misconduct [12]. In this sense, ‘being
replicated’ is probably more common in psychology
than many other sciences because of this explicit
concern (now displayed in HCI) for the lack of actual
replicated studies (or those ‘seen as’ validly replicated).
          </p>
          <p>Psychology’s own debates around its status as a
science are also consonant with these foundational
concerns of ‘being replicable’, and in the replication
agenda we see HCI grasping towards key
epistemological themes which arise in the natural
sciences: alongside ‘observation’, ‘measurement’,
‘description’ and ‘reasoning’ is, of course, ‘replication’.</p>
          <p>If we take HCI as a scientific endeavour (e.g., [11])
then it follows that its concern for replication would
thus be informed by this particular picture of ‘normal
science’; or ‘doing what scientists do’. However, this
assumes coherence of ‘science’ as monolithic practice
as well as mythologising that practice.</p>
          <p>In contrast, ‘being replicated’ is a more pragmatic
question, which concerns what we can learn from
replications and, for example, whether it would be
worthwhile to publish more papers based on replication.</p>
          <p>In order to focus the discussion of replication in HCI, it
would be very helpful if one could gather more
examples from different disciplines, from biology to
physics, to see whether and how replications are valued
in these. Thus we hazard a conjecture: that replication
enjoys a special status within psychology (and the
debate of replication in HCI is thus a reflection of the
influence of psychology, rather than, say, biology, in
HCI). But why might that be?
One issue is with the scale of the question to be
answered through experiment. Some sciences tackle
very detailed and small questions through extremely
detailed experiments. In other words, there exist a very
tight relationship between the data gathered through
the experiments and the derived conclusions. Other
sciences (e.g., social science) tackle bigger questions
and consequently involve a looser relation between
data and conclusion.</p>
          <p>We would argue that there is a ‘scale’ tension in
psychology—and thus HCI—between tackling ‘big’ and
‘minute’ questions, questions that can, or can’t be
settled through experiments. One possible reason for
more replication in psychology is that studies can be
questioned more (i.e., findings are more open to
interpretation).</p>
          <p>Discussion
We have raised some broad issues in the relationship
between replication and HCI, and informed this debate
through recourse to existing work in STS that has
explored replication in the natural sciences.</p>
          <p>
            Firstly we argue for the importance of the increased
consultation of literatures normally foreign to HCI such
as that of STS. This is particularly the case for
situations where knowledge within the field is out of
step with more recent advances in understandings of
scientific knowledge. For instance, our discussions on
replication (and science) within HCI are largely
Popperian or pre-Popperian in form, such as appeals to
ideals such as falsificationism. While we would not
argue against such ideals, we contend that
understanding benefits from expansion, thus as well as
citing Collins, we might also refer to developments by
Kuhn, Feyerabend or Lynch that, for instance,
encapsulate empirical investigations into practical
mundane scientific action [
            <xref ref-type="bibr" rid="ref8">7</xref>
            ].
          </p>
          <p>A fundamental question for the desire for replication in
HCI is that of the motivation to perform replication in
the first place. We need to ask ourselves why we might
bother with replication in the first place and whether
there is any value gained from pursuing a replication
agenda as a distinctive activity within HCI (which is the
position of the workshop call [11]). As we have seen
from STS literature, if we feel the need to derive HCI’s
programme from the methods and epistemological
topics of the natural sciences (e.g., via psychology),
then we must do so knowingly in light of findings from
STS. Thus we argue for different understandings of
replication: a) as an unstable and negotiated practice;
b) as a highly motivated activity rather than as an end
of itself; and c) as playing an important role in the
resolution of scientific controversies. Moving forwards
we would draw attention to the judicious motivated
application of replication—and the need for ‘just why’
and ‘just how’ it is to be pursued. So, we must be clear
about the purposes and motivations of any given
replication beyond abstractly “validating and
understanding contributions” [11].</p>
          <p>Finally, we have argued that a mythological view of
science tends to be implicit in HCI regarding its status
as scientific. This leads us to question the value in
positioning HCI as a scientific endeavour. Thus we
recommend that it would be helpful to separate the
‘foundational’ question (whether HCI is a science) from
the above ‘pragmatic’ question (about the specific
benefits of replication for HCI).</p>
          <p>Acknowledgements
This work is supported by Horizon Digital Economy
Research, RCUK grant EP/G065802/1.</p>
          <p>RepliPRI: Challenges in Replicating
Studies of Online Privacy
Sameer Patil
Helsinki Institute for
Information Technology HIIT
Aalto University
Aalto 00076, FInland
sameer.patil@hiit.fi
Presented at RepliCHI2013. Copyright c 2013 for the
individual papers by the papers authors. Copying permitted only
for private and academic purposes. This volume is published
and copyrighted by its editors.</p>
          <p>Abstract
Replication of prior results has recently attracted
attention and interest from the CHI community. This
paper focuses on the challenges and issues faced in
carrying out meaningful and valid replications of HCI
studies. I attribute these challenges to two main
underlying factors: (i) a domain of inquiry that
simultaneously covers people, social systems, and
technology; and (ii) deficiencies in result reporting and
data archiving. Using examples from investigations of
online privacy, I outline how these challenges manifest
themselves in HCI studies. Longitudinal approaches,
international collaboration, and sharing of study
instruments could help address these challenges.</p>
          <p>Author Keywords
Replication, Privacy, Cultural differences
ACM Classification Keywords
H.1.2 [User/Machine Systems]: Human factors.</p>
          <p>General Terms
Human Factors, Security
Introduction
Replication of prior results has recently attracted
attention and interest from the CHI community. The
resulting discussions tackle replication from two
important perspectives: higher level epistemological
debate on the place and merits of replication in the
scientific (publishing) enterprise and the lower-level
practical considerations for replicating previous studies
from the literature. Growing interest in RepliCHI
suggests increasing recognition for the value of
replicating prior studies. I hope and anticipate that this
trend will foster continued community discussion on
how to justify, appreciate, and reward replication as a
valuable scientific pursuit. Therefore, in this paper I
focus on the latter aspect, viz., challenges and issues
faced in carrying out meaningful and valid replications
of HCI studies.</p>
          <p>I attribute these challenges to two main factors:
1. Domain of inquiry: A large proportion of HCI
studies tackle research problems where results
typically exhibit simultaneous and interacting
influence of individuals, social systems, and
technology. Each of these three factors changes
at drastically different rates and magnitudes. For
instance, technology used in a study may
become obsolete within months or a couple of
years, while physical and cognitive capabilities of
adults change at much slower rates (and the
magnitude of the change is often comparatively
small and predictable). These differences in the
evolution trajectories of humans, cultures, and
technology make it difficult to replicate studies at
a later time and to determine and attribute
causes behind differences in results, if any.
2. Insufficient and/or incomplete reporting: Typically
the only resource available for replicating a study
is the publication describing the results of the
study. Unfortunately, due to page limits and other
editorial reasons, publications often do not
include all information — about methods and/or
data — necessary for carrying out the study the
way it was originally conducted. For instance,
instead of including the entire questionnaire
instrument, the publication may include only
those questionnaire items that led to statistically
significant results. Similarly, results may be
presented in the aggregate or as percentages,
making it difficult to replicate analyses that
require details of individual data points.</p>
          <p>In the following section, I outline how I have found
these challenges to manifest themselves in
investigation of user preferences and practices
regarding online privacy. I conclude with some
thoughts on addressing the challenges.</p>
          <p>
            Replicating Studies of Online Privacy
When thinking about and carrying out replications of
research related to privacy, I have encountered several
practical challenges:
Privacy is a nuanced and complex issue affected by
individual characteristics, context of operation, and the
technology under consideration. For instance,
individuals have been classified into different groups
based on their inherent level of privacy concern [
            <xref ref-type="bibr" rid="ref8">7</xref>
            ],
and privacy concerns have been shown to exhibit
cultural variation [
            <xref ref-type="bibr" rid="ref4">3</xref>
            ]. People’s mental models and
understanding of the underlying technology also affects
their preferences and practices regarding privacy [
            <xref ref-type="bibr" rid="ref5">4</xref>
            ].
          </p>
          <p>This implies that even when considering the same
technology, replication conducted at a later time ought
to take into account the impact of learning effects on
privacy issues. Replications may also encounter the
selection-maturation threat to validity owing to major
external events that occur after the original study, such
as news coverage of privacy breaches. Such events
affect the population’s overall understanding and
awareness of privacy issues, thereby potentially
affecting the results of replications of studies that were
originally conducted prior to these event(s).</p>
          <p>
            The majority of attention in replication has been
devoted to replication at a different (later) time. In the
case of privacy, however, it is equally important to
consider replication across different cultures. For
example, we administered a questionnaire
simultaneously in the US and India, enabling us to
draw interesting and surprising observations from
comparison across cultures [
            <xref ref-type="bibr" rid="ref6">5</xref>
            ]. Our results confirmed
earlier findings regarding low levels of consumer
privacy concerns in India. Surprisingly, by examining
interpersonal privacy separately from consumer
privacy, we found that interpersonal privacy concerns
in India were not only higher than consumer privacy
concerns but also higher than interpersonal privacy
concerns in the US. Our study considered culture at
the broad level of national cultures. However, it should
be noted that for replication purposes “culture” could
be construed to connote any large groups with shared
characteristics and/or values, such as students,
engineers, mothers, liberals, etc. Moreover, if
replication across cultures is conducted at a time later
than the original study, then learning effects and
maturation threats need to be taken into account (as
discussed above).
          </p>
          <p>
            In theory, replication with a different cultural sample is
a simple case of re-running the study with subjects
drawn from a different culture, with translation of
instruments and study materials, if necessary. In
practice, however, cultural differences pose several
hurdles. For instance, the same word or term may be
interpreted differently leading to the same question
being answered differently. For example, we found that
the term “cubicle” was understood differently in the US
and India owing to differences in office layouts and
density. This difference was one of the factors crucial
for understanding the differences in results between
the US and India [
            <xref ref-type="bibr" rid="ref6">5</xref>
            ]. In other studies, I discovered that
the demographic question about ethnicity, which is
commonly asked in the US (and even mandated for
NSF-sponsored studies), was considered potentially
offensive and confusing in Europe. Differences in
lifestyle and beliefs can also affect whether questions
and tasks from one study can yield valid results, or
even make sense, when replicated in a different
cultural context. For instance, some privacy studies
have asked Western respondents about premarital sex,
sexual practices, extramarital affairs, and number of
sexual partners (e.g., [
            <xref ref-type="bibr" rid="ref1">1</xref>
            ]). Such questions are unlikely
to produce meaningful results in cultures where such
practices are uncommon and/or forbidden. Resolving
this issue can be complicated when such
culturally-specific questions comprise parts of standard
scales; using the scale without modifications will not
yield meaningful results and dropping and/or modifying
items in the scale risks affecting the validity of
comparison across studies. Finally, it is also necessary
to consider whether results across cultures are
affected by differences in sampling techniques and
sample characteristics. For instance, although our
comparison of the US and India was limited to software
professionals, the mean and median ages of the Indian
participants were lower than those of the US
participants.
          </p>
          <p>
            We found that understanding privacy-related cultural
nuance often requires insights derived from qualitative
methods (such as interviews, focus groups, field visits,
etc.) and/or insider knowledge of the culture and its
practices [
            <xref ref-type="bibr" rid="ref7">6</xref>
            ]. Currently the CHI community is focused
mostly on replication of studies that employ quantitative
methods, such as experiments, questionnaires, or
usability evaluations. Complementing quantitative
replications with qualitative insights has potential to
broaden the scope of these replication endeavors.
          </p>
          <p>Toward this end, it may also be fruitful to tackle whether
and how qualitative studies could be effectively
replicated.</p>
          <p>Discussion and Conclusion
The previous section utilized examples from
investigations of online privacy attitudes and behaviors
to illustrate some of the challenges and issues in
replicating HCI studies. Online privacy cuts across the
individual, the social, and the technical, in much the
same way as many studies in HCI do. Therefore, I
believe that many, if not all, of these concerns are also
likely to arise in HCI investigations of other topics.</p>
          <p>The RepliCHI workshop is an important milestone
toward developing a comprehensive compilation and
understanding of various challenges involved in the
replication of HCI studies. Moving forward, it is
necessary to apply this knowledge and insight for
constructing best practices to follow and pitfalls to
avoid. Toward this end, I offer suggestions that address
the two important considerations outlined in the
Introduction, viz., (i) domain of inquiry that
simultaneously covers individuals, social systems, and
technology; and (ii) result reporting and data archiving.</p>
          <p>The second of these, in particular, could be easily
addressed by requiring inclusion of full instruments and
study protocols as appendices1. Similarly, authors of
accepted papers could be asked, or even required, to
upload the raw data after taking steps necessary to
protect participant anonymity. In this regard, ACM,
IEEE, NSF, and other prominent HCI funding and
sponsoring organizations can follow the lead of the
NIH, which mandates raw data availability. In a similar
vein, an open source inspired approach could
encourage authors to release the source code of
systems and scripts used for conducting studies and
carrying out analyses. An open question regarding
data and code sharing is how to deal with
commercialization and intellectual property issues
(especially when corporate entities are involved in
conducting the study)2.</p>
          <p>One approach for addressing the issue of intersection
of people and technology is to encourage longitudinal
investigations carried out at regular intervals over
several years. Depending on the details and logistics of
the study, a longitudinal investigation could utilize the
same participants or different participants with the
same sampling method and sample characteristics.</p>
          <p>The former approach can help examine the impact of
changes in individual characteristics, evolution in
lifestyles, and effects of learning. The latter approach
can help illuminate the impact of changes in</p>
          <p>1This also provides the additional benefit of addressing one of the
most common comments raised in peer reviews — lack of
methodological detail.</p>
          <p>
            2Data used by studies conducted by corporations was a hotly
debated topic at the WWW 2012 conference [
            <xref ref-type="bibr" rid="ref2 ref3">2</xref>
            ].
technology. For replications across cultures, however, it
is perhaps best to target simultaneous study
deployment. Fostering international collaborations
and/or leveraging international students to gain cultural
knowledge and access could help in this regard.
          </p>
          <p>Requiring a replication component in Bachelor’s and
Master’s theses could provide a starting point for
repeating studies from the literature, simultaneously
serving a valuable pedagogical purpose by training the
next generation. Further, conferences and journals
could explicitly solicit replications of specific studies.</p>
          <p>Special conference sessions or journal sections could
be devoted solely to replication studies. Discussions
and follow-up activities from the RepliCHI workshop
could lead the way toward legitimizing and promoting
replication as a valuable scientific pursuit within HCI.</p>
          <p>Acknowledgments
I thank Mihir Mahajan and John McCurley for editorial
comments.
researchers.
http://www.nytimes.com/2012/05/22/science/bigdata-troves-stay-forbidden-to-social-scientists.html,</p>
          <p>
            May 2012.
[
            <xref ref-type="bibr" rid="ref4">3</xref>
            ] Milberg, S., Burke, S., Smith, H., and Kallman, E.
          </p>
          <p>Values, personal information privacy, and
regulatory approaches. Communications of the</p>
          <p>
            ACM 38, 12 (1995), 65–74.
[
            <xref ref-type="bibr" rid="ref5">4</xref>
            ] Patil, S., and Kobsa, A. Uncovering privacy
attitudes and practices in Instant Messaging. In
Proceedings of the 2005 International ACM
SIGGROUP Conference on Supporting Group
Work, GROUP ‘05, ACM (New York, NY, USA,
2005), 109–112.
[
            <xref ref-type="bibr" rid="ref6">5</xref>
            ] Patil, S., Kobsa, A., John, A., and Seligmann, D.
          </p>
          <p>
            Comparing privacy attitudes of knowledge workers
in the U.S. and India. In Proceedings of the 3rd
International Conference on Intercultural
Collaboration, ICIC ‘10, ACM (New York, NY, USA,
2010), 141–150.
[
            <xref ref-type="bibr" rid="ref7">6</xref>
            ] Patil, S., Kobsa, A., John, A., and Seligmann, D.
          </p>
          <p>
            Methodological reflections on a field study of a
globally distributed software project. Information
and Software Technology 53, 9 (2011), 969–980.
[
            <xref ref-type="bibr" rid="ref8">7</xref>
            ] Taylor, H. Most people are “privacy pragmatists”
who, while concerned about privacy, will sometimes
trade it off for other benefits. The Harris Poll 17
(2003), 19.
          </p>
          <p>Replicating an International Survey on
User Experience: Challenges,</p>
          <p>Successes and Limitations
Carine Lallemand
Public Research Centre Henri Tudor
29 avenue John F. Kennedy
L-1855 Luxembourg
Carine.Lallemand@tudor.lu
Vincent Koenig
EMACS Research Unit &amp;
Interdisciplinary Centre for
Security, Reliability and Trust
University of Luxembourg
Route de Diekirch
Walferdange, L-7220
Luxembourg
Vincent.koenig@uni.lu
Guillaume Gronier
Public Research Centre Henri Tudor
29 avenue John F. Kennedy
L-1855 Luxembourg
Guillaume.Gronier@tudor.lu
Presented at RepliCHI2013. Copyright © 2013 for the individual papers
byt hep apers’ authors.C opyingp ermittedo nly forp rivatea nda cademic
purposes. This volume is published and copyrighted by its editors.</p>
          <p>
            Abstract
In order to study how the notion of User Experience
(UX) evolved over the last few years, an international
survey originally conducted in 2008 by Law et al. [
            <xref ref-type="bibr" rid="ref1">1</xref>
            ]
has been replicated. Its main goal was to get some
insights on the points of view from practitioners on the
notion of UX. After having slightly adapted the initial
(English) survey and having translated it into French
and German, more than 758 valid answers have been
collected from all over the world. This experience report
aims at illustrating some of the challenges involved in
the replication of such a study as well as successes and
limitations.
          </p>
          <p>Author Keywords
User Experience; HCI Research; Replication; Survey;
Experience Report
ACM Classification Keywords
H.5.m. Information interfaces and presentation (e.g.,
HCI): Miscellaneous.</p>
          <p>
            General Terms
Human Factors; Design; Measurement
Introduction: a Tale of Two Studies
Some concepts in the field of HCI are widely spread and
used by practitioners even if a lack of empirical
research prevents the true understanding of their
meaning and impacts [
            <xref ref-type="bibr" rid="ref1 ref2 ref3 ref4">3, 1, 2</xref>
            ]. This is the case for User
Experience (UX). Despite many attempts to
understand, define and scope UX, it is still not clear
whether a consensus has been reached on this concept
or not. In a willingness to address the complexity of the
UX concept, to contribute to its further development
and consolidation, we decided to replicate a previous
survey entitled “Understanding, scoping and defining 
UX: a surveya pproach” [
            <xref ref-type="bibr" rid="ref1">1</xref>
            ].
          </p>
          <p>The original study has been first spread during the
main conference CHI’08 before being broadcast through 
several communication channels. Results have been
published the following year in the proceedings of
CHI’09, as a 10-pages long paper. 275 answers had
been collected at that time from 25 countries.</p>
          <p>In order to adapt to our project’s multicultural context 
and to reach a wider audience within the
Frenchspeaking community of UX practitioners, all
questionnaire items have been translated, from the
English master version to French and German (both
languages being commonly used in Luxembourg). A
back translation process has been applied to ensure the
quality and validity of the process.</p>
          <p>Rationale for a Replication
Several reasons may explain the choice to replicate this
UX survey. First of all, as User Experience is still a
concept in maturation, it was worth taking stock of the
situation four years after the initial study in order to
see a possible evolution in the representations, points
of view and practices associated to UX. Replication acts
here as a way to check whether the results still apply in
a different context to the original study, especially in a
different temporality.</p>
          <p>Moreover, the translation into two others languages
allowed us to reach a wider audience, especially in the
multicultural context in which the present work was
involved. As this study constituted an exploratory step
within a wider Luxemburgish project focused on UX
Design, gathering additional knowledge about the
French- and German-speaking practitioners’ community 
(not well represented in the initial study) seemed
crucial to us. By trying to draw an accurate picture of
the current situation of UX and building on that basis,
we aim at achieving the best solutions possible to
design for UX.</p>
          <p>Form of Replication
This study may be considered as a direct replication,
since differences between both studies are limited to:
 A minor extension through the translation in French
and German languages. The original English
version was kept as default language and still
represented 58.4 % of the completed surveys.
 Additional sociodemographics items aimed at better
categorizing participants and acting as control
variables to analyze the data.</p>
          <p>
            Summary of the Methodology
Structure of the Survey
The UX questionnaire encompasses 3 sections:
 Background: respondents were asked to first
answer 13 questions about their job and
educational background, their level of familiarity
with UX or the importance of UX in their actual
interesting and relevant information. Much research has
been done on modifying search ranking based on social
signals for web pages [
            <xref ref-type="bibr" rid="ref1">1</xref>
            ][
            <xref ref-type="bibr" rid="ref2 ref3">2</xref>
            ][
            <xref ref-type="bibr" rid="ref4">3</xref>
            ][
            <xref ref-type="bibr" rid="ref6">5</xref>
            ][
            <xref ref-type="bibr" rid="ref7">6</xref>
            ], but how should
we present the social signals for web search results?
The most recent paper that we have found is the
CHI2012 paper on social annotations by Muralidharan
et al. [
            <xref ref-type="bibr" rid="ref5">4</xref>
            ].
          </p>
          <p>
            Previous Research
Muralidharan et al. [
            <xref ref-type="bibr" rid="ref5">4</xref>
            ] studied the perception of social
annotations appearing below search results, as in Figure
1. Consistent with prior papers, we use the term “social
signals” to refer to any social information that is used
to affect ranking, recommendation or presentation to
the user. We use the term “social annotations” to refer
to the presentation of social signals for an explanation
as to why a search or recommendation result is
presented. Thus, a social signal only becomes an
annotation when it is presented to the user.
          </p>
          <p>Study Protocol
Their first study had two parts: (1) In the first part,
participants conducted 18-20 search tasks, randomly
ordered. Half were designed so that one or two social
annotations would appear in the top four or five results.</p>
          <p>The search results pages were presented as static mocks
that were generated before the study, customized for
each participant.
(2) The second part consisted of a retrospective
thinkaloud (RTA) where they walked the participant through
each task using the eyetrace data post hoc. During the
interview, researchers checked noticeability by asking if
the participants noticed the social annotations, either by
them mentioning they saw them or being explicitly asked
if they had seen them. During the RTA the researchers
also obtained qualitative feedback about social
annotations.</p>
          <p>The second study compared the perception of multiple
designs of social annotations. They varied profile image
size (small, large), snippet length (1, 2, 4 lines), and
annotation position (above, below snippet). For this study
the same mocks were used for each participant, with
customization only for customizing familiar names and
faces of people in the annotations. In the second study,
noticeability of the annotations was measured by counting
the number of fixations.</p>
          <p>Findings
In the first study, they found that only 5 of the 45 (11%)
of the visible social annotations were noticed. In the
second study, they found that there were fewer fixations
on annotations when: the snippet length was longer; the
image was smaller; and the annotation was below the
snippet. They concluded that the optimal design for a
social annotation is one with a large picture, above the
snippet, with a short snippet length.</p>
          <p>Our Method and Replication
We aimed to actually test the proposed annotation
design guidelines from study 2 using live user data to
see if people notice the annotations more by using the
method from study 1. Specifically, we wanted to test
with live data that is relevant to participants (from their
connections), as opposed to the static images used
previously. An example of a social annotation with the
new design is shown in Figure 2.</p>
          <p>Study Protocol
Experimental sessions consisted of 3 parts, the first two
using essentially the same protocol as experiment 1’s in
the previous work, with some improvements.</p>
          <p>PART 1: SEARCH TASKS
We designed planned 16-20 custom search tasks for
each subject, at least eight of which were “social
search" tasks designed to organically pull up social
annotations. The 8 non-social search tasks were the
same as used in the prior work.</p>
          <p>In order to ensure that personal results appear for as
many queries as required, we designed 2-4 additional
social search tasks for each participant that were
intended to bring up personal results. This way, if one
social search task did not bring up personal results, we
gave them the additional tasks to help ensure that they
saw 8 tasks with personal results.</p>
          <p>PART 2: RETROSPECTIVE THINK-ALOUD
After the search tasks, we immediately conducted a
retrospective review of eye-tracking traces for search
tasks in which subjects exhibited behaviors of some
interest to the experimenter. Review of eye-tracking
videos prompted think-aloud question answering about
participants’ process on the entire task, particular
interesting pages, and particular interesting results.</p>
          <p>
            Unlike Experiment 1 in Muralidharan et al. [
            <xref ref-type="bibr" rid="ref5">4</xref>
            ], we
examined the eyetrace data directly by hand to
determine noticeability, rather than through verbal
feedback during the RTA tasks.
          </p>
          <p>PART 3: THINK-ALOUD TASKS
Finally, participants performed two or three different
search queries for which we determined ahead of time
that should bring up relevant personal results. Here we
gathered qualitative feedback on social annotations.</p>
          <p>Results
In total, we collected eye-trace data for 153 tasks from
nine subjects. Each eye-trace data for each task was
analyzed by hand by an experimenter to understand:
which positions contained personal search results;
whether the search result was in the field of view in the
browser; and importantly, whether the subject fixated
on the result and/or the social annotation. This funnel
analysis approach is different than the previous work’s
approach of asking participants if they noticed the
annotations.</p>
          <p>
            We discovered that participants fixated on annotations
in 35 of the 58 tasks where they appeared (60%). This
is a dramatic improvement over the 11% perception
rate of the Muralidharan et al. [
            <xref ref-type="bibr" rid="ref5">4</xref>
            ]. We account this
difference primarily to the new annotation design.
          </p>
          <p>Replication Discussion
Access to Previous Experimental Data. We were able to
repeat the exact same tasks performed in the previous
work but only because we share a co-author who had
access to the data. If anyone else tried to replicate the
study, they would not have been able to do so as
effectively.</p>
          <p>
            Temporal Challenges. Even though the search tasks
were identical, because the study was conducted
several months later, some of the task questions were
no longer topically relevant. For example, one task
asked “What is the website for the Google image
labeling game?” At the time of our study, the website
was no longer active. Similarly, the search task “Find
some information about the Nevada law legalizing
selfdriving cars” brought up news articles from the
previous summer, when Muralidharan et al. [
            <xref ref-type="bibr" rid="ref5">4</xref>
            ]
conducted their research, since it was no longer recent
news.
          </p>
          <p>This raises a big issue for research replication:
changing environments such as time or space. In our
case, the tasks lost their relevancy over time.</p>
          <p>Researchers could help mitigate this by rewriting tasks
so they are more relevant but still in the same vein as
the original. For example, we could have written a
different task that was more topical but would still be
categorized as news. It must be decided which would
cause the least amount of discrepancy for replication:
maintaining the identical, less relevant task or rewriting
a relevant task that differs from the original.</p>
          <p>Iteration and Refinement. The primary difference in our
protocol, measuring perception with fixation data rather
than verbal confirmation, offered an improvement to
the previous work.</p>
          <p>
            Even with those challenges, we feel that we were
successful in our replication efforts. We conducted an
almost identical study to confirm the proposed
improved design for social annotations and found a
large increase in perception.
[
            <xref ref-type="bibr" rid="ref5">4</xref>
            ] Muralidharan, A., Gyongyi, Z., and Chi, E. H. Social
annotations in web search. In Proc. CHI 2012, 1085–
1094.
[
            <xref ref-type="bibr" rid="ref6">5</xref>
            ] Yanbe, Y., Jatowt, A., Nakamura, S., and Tanaka,
K. Can social bookmarking enhance search in the web?
In Proc. JCDL 2007, 107–116.
[
            <xref ref-type="bibr" rid="ref7">6</xref>
            ] Zanardi, V., and Capra, L. Social ranking:
uncovering relevant content using tag-based
recommender systems. In Proc. RecSys 2008, 51–58.
          </p>
          <p>Quincy Brown Germaine Irwin
Games+Mobile Play Learn Live Lab UMBC
Bowie State University Information Systems
14000 Jericho Park Road 1000 Hilltop Circle
Computer Science Building Baltimore, MD 21250 USA
Bowie, MD 20715 USA germaine.irwin@umbc.edu
qbrown@bowiestate.edu
Lisa Anthony
UMBC
Information Systems
1000 Hilltop Circle
Baltimore, MD 21250 USA
lanthony@umbc.edu
Robin Brewer
UMBC
Information Systems
1000 Hilltop Circle
Baltimore, MD 21250 USA
brewer3@umbc.edu</p>
          <p>Challenges of Replicating Empirical
Studies with Children in HCI</p>
          <p>General Terms
Design, Human Factors.</p>
          <p>Abstract
In this paper, we discuss the challenges of conducting a
direct replication of a series of mobile device usability
studies that were originally conducted with adults and
older children (ages 7 to 17). The original studies were
designed to investigate differences in how adults and
children use mobile devices to touch targets and create
Jaye Nias surface gestures. In this paper, we report on a
Games+Mobile Play Learn Live Lab replication we conducted with young children (ages 5 to
Bowie State University 7). We discuss several methodological changes that
14000 Jericho Park Road were needed to elicit the same quality of data from the
Computer Science Building replication with young children as had been obtained
Bowie, MD 20715 USA from the older children and adults. The insights we
jayeaclark@aol.com present are relevant to the extension of empirical</p>
          <p>studies in HCI in general to younger children.</p>
          <p>Berthel Tate
Games+Mobile Play Learn Live Lab
Bowie State University
14000 Jericho Park Road
Computer Science Building
Bowie, MD 20715 USA
TATEB0528@students.bowiestate.
edu</p>
          <p>Author Keywords
Child-computer interaction, touch interaction, gesture
interaction, mobile devices, replication, empirical study.</p>
          <p>ACM Classification Keywords
H.5.m. Information interfaces and presentation (e.g.,
HCI): Miscellaneous.</p>
          <p>
            Introduction
In the context of research studies, children have often
been viewed as vulnerable or ill-equipped and have
been excluded from participation in studies due to
concerns regarding informed consent, confidentiality,
and the specialized attention or procedures required
when conducting research with minors [
            <xref ref-type="bibr" rid="ref1 ref2 ref3">1, 2</xref>
            ]. To that
end, previous research in human-computer interaction
(HCI) has focused on including child participants by
developing child-centered research methods and
adapting protocols specifically for children [
            <xref ref-type="bibr" rid="ref1 ref4">1, 3</xref>
            ].
          </p>
          <p>Though these efforts have increased the inclusion of
children in HCI research, the use of separate protocols
does not always allow for the direct comparison of
findings between adults and children.</p>
          <p>
            Here we present our insights from a direct replication of
a series of studies of touch and gesture interaction on
mobile devices that was first conducted with adults (18
years and older) and older children (ages 7 to 17) [
            <xref ref-type="bibr" rid="ref7 ref8 ref9">6,
7, 8</xref>
            ], and then replicated with younger children (ages 5
to 7). The goal of replicating these studies with younger
children was to evaluate whether the same findings for
older children and adults would hold for younger
children. Though the studies were previously conducted
with children as young as 7 years old, evidence from
developmental psychology literature prompted us to
include even younger participants: typically, as
individuals mature from early childhood to adulthood,
their cognitive and physical abilities also mature [
            <xref ref-type="bibr" rid="ref5 ref6">4, 5</xref>
            ].
          </p>
          <p>Thus, the inclusion of younger children will allow for the
comparison of patterns across all age groups, and
support our overall goal of helping mobile application
developers create more age-appropriate apps for
children vs. adults, or even universally accessible apps.</p>
          <p>
            Original Study Design
We have previously conducted three studies with adults
(over 18 years old) and children (ages 7 to 17) [
            <xref ref-type="bibr" rid="ref7 ref8 ref9">6, 7,
8</xref>
            ] to investigate mobile device input and interaction
differences between adults and children. The
applications we used were designed specifically for
these studies. Each participant completed a gesture
task and target task. For the gesture tasks, participants
used their finger to draw gestures (i.e., letters,
numbers, symbols, and shapes) on the device screen.
          </p>
          <p>For the target tasks, participants touched square
targets on the phone screen. A summary of the tasks
from each study is given in Table 1. We also describe
each task briefly to highlight the key points.</p>
          <p>[P6re]lim. [S7tu,d8y] 1 [S8tu]dy 2 Replica
No. Kids 8 16 25 7
(Ages) (7 to 11) (7 to 16) (10 to 17) (5 to 7)
No.</p>
          <p>Adults 6 14 16 N/A
(18+)
No FB &amp; FB No FB &amp;</p>
          <p>Gesture FB Gesture</p>
          <p>Table 1. Tasks and Studies.</p>
          <p>
            Mini Target Task [
            <xref ref-type="bibr" rid="ref7">6</xref>
            ]
Square targets (43 in all) of four different sizes, large
(26.4mm), medium (15.8mm), small (10.5mm) and
very small (5.29mm), were displayed to the user one at
a time. As the participant attempted to touch a target,
the application logged the touch event. Participants
were allowed one attempt per target only; touches
were scored as hits or misses.
          </p>
          <p>
            Target Task [
            <xref ref-type="bibr" rid="ref8 ref9">7, 8</xref>
            ]
The full target task used 104 targets of 4 different
sizes: very small (3.175 mm), small (6.35 mm),
medium (9.5 mm), and large (12.7 mm), in 13
different interface positions. This task incorporated
edge padding for half the targets, which caused them
to appear close to, but not on, the edge of the screen.
          </p>
          <p>The order of targets was designed to evenly represent
all possible transitions between target positions and
sizes, and no two consecutive targets had the same
size or position. Unlike the mini target task, to advance
to the next target, the participants had to successfully
touch within the boundaries of the visible target.</p>
          <p>Therefore, multiple attempts for the same target were
possible; touches were again scored as hits or misses.</p>
          <p>
            Gesture Task – Feedback [
            <xref ref-type="bibr" rid="ref8 ref9">7, 8</xref>
            ]
Participants were shown a screen with text indicating
which gesture to make and a “Done” button. Users
used their finger to draw gestures on the device screen
and press “Done” when finished. The complete gesture
set (20 in all) included letters (A, E, K, Q, and X),
numbers (2, 4, 5, 7, and 8), symbols (line, plus, arch,
arrowhead, and checkmark), and geometric shapes
(circle, square/rectangle, triangle, diamond, and heart).
          </p>
          <p>Participants were given a paper sheet showing what
each gesture should look like, in case they were not
familiar with every symbol by name (especially relevant
for children). Participants entered an example of each
gesture type one after another, and repeated this five
times, yielding a total of six examples of each gesture
type. As participants drew each gesture, a trace
appeared under their finger of the gesture, but they
were not able to edit their gestures.</p>
          <p>
            Gesture Task – No Feedback [
            <xref ref-type="bibr" rid="ref7 ref9">6, 8</xref>
            ]
The no feedback gesture task was identical to the
feedback task except participants did not see a trace of
the symbol beneath their finger as they drew.
          </p>
          <p>The Replica
We replicated Study 2 (conducted with older children
and adults, see Table 1) using the same task
applications: participants in the replicated study
completed the Gesture Task – No Feedback, Gesture
Task – Feedback, and the Target Task. So far, we have
had 7 participants in this replication; three were 5
years old, one was 6 years old, and three were 7 years
old. Of these participants, four were females, one
participant was left-handed, and most self-rated their
familiarity with touch input devices to be “average.”
Successes of Replica
The primary aspect of the protocol from the original
study that was successful was the Target Task: in
general, the 5 to 7 year olds were able to complete the
Target Task without much difficulty. We believe this
was because this task is very short and takes little time
(about 1 to 2 minutes) compared to what is required to
complete the six iterations of the gesture task (about 8
to 10 minutes). Furthermore, the Target Task required
participants to perform an action (touching the
interface) with which most children were familiar. In
contrast, most of the children were not familiar with all
of the gestures they had to draw in the Gesture Tasks
and had to practice creating the gestures.</p>
          <p>Limitations of the Replication
While the Target Task was a success, we encountered
problems with the younger participants not completing
all repetitions in the Gesture Tasks. Only 2 of 7 children
completed all iterations of the Feedback and No
Feedback Gesture Tasks. The average number of
rounds completed was less than 3 for the other
children. With the majority completing so little of the
task, we did not have enough data to be confident in
results from a gesture recognizer (which needs enough
data for both a training set and a testing set).</p>
          <p>
            Comparison of Results from the Replica
Target Task – Misses
Table 2 shows the proportion of targets missed on the
first attempt in the Target Task for all prior studies [
            <xref ref-type="bibr" rid="ref7 ref8 ref9">6,
7, 8</xref>
            ] and the replica with younger children. The 34%
miss rate for the replica is higher than the Study 1 [
            <xref ref-type="bibr" rid="ref8 ref9">7,
8</xref>
            ] and Study 2 [
            <xref ref-type="bibr" rid="ref9">8</xref>
            ] miss rate, which we hypothesize is
due to the younger age of the participants (the
preliminary study [
            <xref ref-type="bibr" rid="ref7">6</xref>
            ] had a higher miss rate because
the task only allowed one attempt per target).
          </p>
          <p>
            Prelim. [
            <xref ref-type="bibr" rid="ref7">6</xref>
            ]
Study 1 [
            <xref ref-type="bibr" rid="ref8 ref9">7, 8</xref>
            ] 17%
Study 2 [
            <xref ref-type="bibr" rid="ref9">8</xref>
            ]
Replica
          </p>
          <p>Adults
32%
16%
N/A</p>
          <p>
            Children
46%
23%
23%
34%
Study 1 [
            <xref ref-type="bibr" rid="ref8 ref9">7, 8</xref>
            ] 90% (FB only)
          </p>
          <p>
            81% (FB only)
Study 2 [
            <xref ref-type="bibr" rid="ref9">8</xref>
            ]
Replica
91% (FB),
91% (no FB)
N/A
82% (FB),
85% (no FB)
46% (FB),
49% (no FB)
          </p>
          <p>
            Gesture Task
Table 3 includes the average per-user recognition
results (computed for the replica using the open-source
$N multistroke recognizer [9], as in prior work [
            <xref ref-type="bibr" rid="ref8 ref9">7, 8</xref>
            ])
for both Gesture Tasks across three of the studies. Both
in spite of, and as a result of, the lower number of
gesture samples collected so far during the replica, the
replicated study recognition results are consistent with
the overall trend we have found in our work that
recognition rates are lower for younger participants. To
ensure this finding is robust, we intend to explore ways
to encourage children to complete the tasks so that we
can examine this trend in more depth for the youngest
children. We also hypothesize that the lower
recognition rates may be attributed to the grade level
of some of the participants (some 5 and 6 year olds
had not completed first grade). Children who had been
to school had more practice with handwriting and made
gestures that appeared be more canonical (Figure 1).
          </p>
          <p>(a)
(b)</p>
          <p>(c)</p>
          <p>Reasons for Accounted Differences
In general, we found that the results of the replica were
consistent with the original studies. However, we have
identified four challenges areas with respect to younger
children not completing the gesture task portion of the
study that could be useful for doing similar empirical
replications with younger children in the future.</p>
          <p>
            Motivation. All participants, adults and children, were
compensated $10 for their participation in Study 1,
Study 2, and the replica [
            <xref ref-type="bibr" rid="ref8 ref9">7, 8</xref>
            ] (the preliminary study
had no compensation [
            <xref ref-type="bibr" rid="ref7">6</xref>
            ]). Though financial
compensation may motivate adults, we noted that the
delayed financial compensation (receiving $10 after the
study vs. immediate rewards throughout the study)
might not have been enough motivation for the young
children in the replica.
          </p>
          <p>
            Attention Span. We also noted that the young
participants of the replica seemed less focused than the
older participants from the original studies [
            <xref ref-type="bibr" rid="ref7 ref8 ref9">6, 7, 8</xref>
            ].
          </p>
          <p>
            For example, they frequently told stories to the
experimenter while completing the tasks, especially
during the Gesture Tasks, and many asked for water or
breaks during the session. Older participants in the
original work did not exhibit this behavior [
            <xref ref-type="bibr" rid="ref7 ref8 ref9">6, 7, 8</xref>
            ].
          </p>
          <p>
            Research Setting. All of the studies were completed in
an academic usability lab with no windows [
            <xref ref-type="bibr" rid="ref7 ref8 ref9">6, 7, 8</xref>
            ].
          </p>
          <p>This setting may not have been inviting and
comfortable for the young participants of the replica. In
the future, we plan to conduct studies in a more
kidfriendly environment, such as a bright room with
natural light and pleasant surroundings.</p>
          <p>
            Acknowledgements
This work was partially supported by Department of
Education HBGI Grant Award #P031B09020 and
National Science Foundation Grant Awards
#IIS1218395 / IIS-1218664. Any opinions, findings, and
conclusions or recommendations expressed in this
paper are those of the authors and do not necessarily
reflect these agencies’ views.
[
            <xref ref-type="bibr" rid="ref4">3</xref>
            ] Druin, A. (1999). Cooperative Inquiry: Developing
New Technologies for Children with Children. Proc. ACM
CHI 1999, 592-599.
[
            <xref ref-type="bibr" rid="ref5">4</xref>
            ] Thomas, J.R. (1980). Acquisition of Motor Skills:
Information Processing Differences Between Children
and Adults. Research Quarterly for Exercise and Sport
51(1), 158–73.
[
            <xref ref-type="bibr" rid="ref6">5</xref>
            ] Piaget, J. (1983). Piaget’s Theory. In P. Mussen,
ed., Handbook of Child Psychology. Wiley &amp; Sons, New
York, NY, USA.
[
            <xref ref-type="bibr" rid="ref7">6</xref>
            ] Brown, Q. and Anthony, L. (2012). Toward
Comparing the Touchscreen Interaction Patterns of Kids
and Adults. ACM CHI 2012 EIST workshop, 4pp.
[
            <xref ref-type="bibr" rid="ref8">7</xref>
            ] Anthony, L., Brown, Q., Nias, J., Tate, B., and
Mohan, S. (2012). Interaction and Recognition
Challenges in Interpreting Children’s Touch and Gesture
Input on Mobile Devices. Proc. ACM ITS 2012, 225-234.
[
            <xref ref-type="bibr" rid="ref9">8</xref>
            ] Anthony, L., Brown, Q., Tate, B., Nias, J., Brewer,
R., and Irwin, G. (In press). Designing Smarter
TouchBased Interfaces for Educational Contexts. Journal of
Personal and Ubiquitous Computing: Special Issue on
Educational Interfaces, Software, and Technology, to
appear.
[9] Anthony, L. and Wobbrock, J. O. (2012).
$NProtractor: A Fast and Accurate Multistroke Recognizer.
          </p>
          <p>In Proc. Graphics Interface 2012, 117-120.</p>
          <p>Replicating Residential Sustainability
Study in Urban India</p>
          <p>Mohit Jain
IBM Research Labs
Bangalore 560045 India
mojain13@in.ibm.com
Yedendra B. Shrinivasan
IBM Research Labs
Bangalore 560045 India
yshriniv@in.ibm.com
Tawanna Dillahunt
School of Information
University of Michigan
4340 North Quad
105 S State Street
Ann Arbor, MI 48109
tdillahu@umich.edu</p>
          <p>Abstract
Despite the global nature of problems such as rapid
depletion of fossil fuels and water resources, most of
the solutions being developed to address these issues
are based on studies done in the developed world. We
conducted a study of energy, water and fuel
conservation practices in urban India, replicating the
work of Dillahunt et al., a qualitative study that
explored the current practices, beliefs and attitudes of
low-income households in two distinct U.S. locations.</p>
          <p>We used the same method, a photo-elicitation interview
study, with 11 participants in Bangalore, India. Our
study highlights deep conservation actions, which were
influenced by the cultural context and different from
the original work. Participants in our study shared
motivations to conserve with participants in the
previous study including scarcity, money, comfort and
religion.</p>
          <p>The purpose of this paper is to shed insight on our
replication study. We discuss the purpose for
conducting the replication study and describe the
procedures we followed; we also provide information
regarding access to procedures and data analysis
techniques used from the original study. We discuss
subtle differences in our procedure and how this may
have affected our results and discuss key findings from
our replication.</p>
          <p>Author Keywords
Energy; Sustainability; Developing World.</p>
          <p>ACM Classification Keywords
H.5.m. Information interfaces and presentation (e.g.,
HCI): Miscellaneous.</p>
          <p>General Terms
Human Factors; Design; Measurement.</p>
          <p>
            Introduction
The goal of our study was to elicit a detailed picture of
consumption and conservation practices and beliefs in
Indian households. Like some of the prior work
conducted in developed nations (primarily in the U.S.
e.g. [
            <xref ref-type="bibr" rid="ref1">1</xref>
            ], [
            <xref ref-type="bibr" rid="ref4">3</xref>
            ], [
            <xref ref-type="bibr" rid="ref8">7</xref>
            ], [
            <xref ref-type="bibr" rid="ref9">8</xref>
            ]), we were interested in
understanding motivations behind the conservation
practices and challenges our participants faced around
resource management. We decided to conduct our
study in a developing nation as there was little
information about whether or how prior results applied
to other geographies, cultures, and socioeconomic
groups. Further, we chose to focus on middle and
highincome households because they consume resources in
more diverse ways (e.g., own multiple types of
appliances). Since our study was exploratory in nature,
we chose to replicate a study conducted to understand
energy consumption among low-income households in
two U.S. locations [
            <xref ref-type="bibr" rid="ref1">1</xref>
            ].
          </p>
          <p>
            Replication
The original study conducted used photo-elicitation
interviews [
            <xref ref-type="bibr" rid="ref1">1</xref>
            ], which produces a different kind of
information provoking feelings and memories. This
information is not as easy to gather using standard
interviewing techniques. Further, pictures provide a
focal point of conversation, which helps to alleviate any
awkwardness an interviewee may feel [
            <xref ref-type="bibr" rid="ref1">1</xref>
            ]. Further,
photo-elicitation interviews make it easy to agree on
categories when analyzing data [
            <xref ref-type="bibr" rid="ref2 ref3">2</xref>
            ].
          </p>
          <p>
            We analyzed the data using the same technique
described in Dillahunt, et al. [
            <xref ref-type="bibr" rid="ref1">1</xref>
            ]. We coded and
analyzed our interview data in an iterative fashion
following methods taken from informed grounded
theory [
            <xref ref-type="bibr" rid="ref7">6</xref>
            ].
          </p>
          <p>Though photo-elicitation interview studies have been
conducted in the past and well documented, in
replicating the original study, we identified some
aspects of the study that needed to be taken into
account across various populations. For example, we
made some changes in the protocol to factor in new
contexts such as cultural differences.</p>
          <p>Next, we discuss our method and differences that may
have affected the results between the two studies.</p>
          <p>Methodology
Prior conducting our study, we contacted the original
researchers for their IRB material. This included
recruiting detail, the surveys used to collect
demographic information, and the specific script
researchers read to participants. We made slight
variations in the survey to accommodate for cultural
context, such as the types of household appliances and
transportation options. For example, we did not include
dryers in our appliance list, as they were not as
common among our population; we also added water
heaters (Geyser) to the list. To understand
conservation behavior we asked questions such as
whether participants left the fan on to dry clothes, use
solar water heaters to heat water, conduct regular
refrigerator maintenance, and/or use inverters (UPS).</p>
          <p>We also removed questions related to religion and
spirituality as few participants were offended or felt
uncomfortable answering those questions (though we
made answering those questions optional). One such
question was if they were motivated to conserve
resources to protect God’s creation. Access to this
information helped in replicating the study method in
its original form.</p>
          <p>Differences in protocol
Despite being able to replicate all aspects of the study,
there were some subtle differences that may have
affected our results. These included the technology
used to capture photos, payment, recruitment and the
type of researchers conducting the study.</p>
          <p>In the original study, participants used disposable
cameras and at least one participant had never used a
camera before the study. Our participants used either a
digital camera or the cameras on their personal phones.</p>
          <p>Our participants had prior experience using the
cameras. With these differences, participants using
their own (digital) cameras may have felt more
comfortable taking pictures and they may have been
less concerned with running out of exposures. Though
this unlikely had an impact on the results, it is a
difference that should be considered.</p>
          <p>The original study compensated participant for the time
they spent during the interview. We had a different
payment model. We did not pay our participants
directly because we found during our interviews that
participants were not interested in receiving payment.</p>
          <p>Instead, we paid our participants 2500INR to a charity
organization for every 50 participants to complete our
online survey (the results of our survey were removed
from our final paper submission).</p>
          <p>The original study was conducted as a university study,
whereas we were industry researchers conducting the
same study. We were studying two distinctly separate
populations, which makes it unclear how this may have
influenced participant attitudes. As both studies were
conducted in participant households, this may have
alleviated any differences participants felt in terms of
how comfortable they were in being interviewed. Our
methods for recruiting were limited because we
conducted our study as a private organization. As a
result, we did not advertise publically—we relied on
word of mouth and snowball sampling, which may have
added bias to our participants.</p>
          <p>From an internal organizational perspective, the “IRB”
process for working with participants is slightly more
difficult than in university settings. Industry is
concerned about privacy issues such as IP; however,
whether or not this is transparent to participants and
affects their attitudes was not well understood.</p>
          <p>
            Results
Many of our participants’ conservation practices and
motivations matched key categories of actions noted in
the original study; however, as expected, the findings
were not identical. We were able to contribute new
categories and also leverage a vocabulary described in
a more recent study, which provided evidence that the
authors’ framework generalized across different
populations and cultures [
            <xref ref-type="bibr" rid="ref4">3</xref>
            ].
          </p>
          <p>
            We also saw how our results generalized with the study
we replicated and past studies of home energy
consumption in developed regions. For example,
participants in our study shared motivations to
conserve with participants in past studies of typical [
            <xref ref-type="bibr" rid="ref1">1</xref>
            ],
[
            <xref ref-type="bibr" rid="ref4">3</xref>
            ] and low-income households [
            <xref ref-type="bibr" rid="ref1">1</xref>
            ] including money,
comfort and religion. Barriers to conservation such as
money, comfort and safety also overlapped past
studies. We highlighted two key differences between
our findings and others in our final paper [
            <xref ref-type="bibr" rid="ref6">5</xref>
            ]. These
include the impact of resource shortages (scarcity) and
the value of eco-feedback.
          </p>
          <p>When looking to generalize across lower-income U.S.
households, our participants did not mention many
common conservation behaviors. Our examples
included re-using plastic drinking bottles for storing oils
instead of buying dedicated containers, packing a
family of 5 or 6 onto a single moped, and washing
dishes using sand, ash, or coconut husk where water is
in short supply—all findings unique to Indian culture.</p>
          <p>However, India has wide socio-economical, cultural,
and demographic diversity, which makes it difficult to
know exactly how broadly these findings generalize
even within the country.</p>
          <p>
            The major reason for differences among our work and
the work replicated [
            <xref ref-type="bibr" rid="ref1">1</xref>
            ] is the shift in the cultural
context. Hence we obtained many conservative actions,
related to the Indian culture, but may not be relevant
for developed countries.
          </p>
          <p>Key Insights
We believe we can offer three key insights from our
replication study. First, having access to scripts that
describe the research method, the surveys conducted,
recruiting material, and access to a responsive original
author, simplified our process. This information is often
available in research Institutional Review Board
documentation (IRBs); however, it is unclear whether
this material is typically shared among researchers.</p>
          <p>Further, we are somewhat limited in our recruiting
efforts due to the rigor required to advertise publically.</p>
          <p>This limited the types of participants that we could
recruit and perhaps biased our results. Nevertheless,
we found similarities between our results and the
original study’s results, as well as similarities between
other home consumption studies.</p>
          <p>Finally, in our study, we found the need to modify our
demographic and baseline survey to account for
cultural differences that existed between our study
population, such as the types of resources used.</p>
          <p>Discussion
Our replication was somewhat atypical as it was a
replication of a qualitative study. However, our aim was
not to replicate prior results. Our study was exploratory
and we expected to see some conflicting results
because of cultural and socioeconomic differences
between the two populations; however, we anticipated
some overlap as well. One topic for discussion is
whether we can truly “replicate” a qualitative study.</p>
          <p>
            What exactly does it mean to replicate a qualitative
study? Another question to consider is if using the
same surveys was limiting in any way? We had to
modify the survey based on cultural differences but was
having the original material as a starting point a
limitation?
[
            <xref ref-type="bibr" rid="ref1">1</xref>
            ] Dillahunt, T., Mankoff, J., Paulos, E., and Fussell, S.
          </p>
          <p>It's not all about “Green”: energy use in low-income
communities. Ubicomp 2009, 255-264.
[9] World Population Data Sheet 2012.
http://www.prb.
org/pdf12/2012-population-datasheet_eng.pdf
Replicating and Applying a Neuro-Cognitive
Experimental Technique in HCI Research
David Coyle
Interaction and Graphics Group,
Dept. of Computer Science,
University of Bristol,
Bristol BS8 1UB, UK
david.coyle@bristol.ac.uk
Presented at RepliCHI2013. Copyright © 2013 for the individual
papers by the papers’ authors. Copying permitted only for private
and academic purposes. This volume is published and copyrighted
by its editors.</p>
          <p>
            Abstract
In cognitive neuroscience the sense of agency is
defined as the as the experience of controlling one’s
own actions and, through this control, affecting the
external world. At CHI 2012 I presented a paper
entitled “I did that! Measuring Users’ Experience of
Agency in their own Actions” [
            <xref ref-type="bibr" rid="ref1">1</xref>
            ]. This extended
abstract draws heavily on that paper, which described
an implicit measure called intentional binding. This
measure, developed by researchers in cognitive
neuroscience, has been shown to provide a robust
implicit measure for the sense of agency. My interest in
intentional binding stemmed from prior HCI literature,
(e.g. the work of Shneiderman) which emphasises the
importance of the sense of control in human-computer
interactions. The key question behind the CHI 2012
paper was: can we apply intention binding to provide
an implicit measure for the experience of control in
human-computer interactions? In investigating this
question, replication was a key element of the
experimental process.
          </p>
          <p>
            Keywords
Replication; intentional binding; the experience of
agency; evaluation methods
ACM Classification Keywords
H.5.2 [Information Interfaces and Presentation]: User
Interfaces - Evaluation/methodology;
Intentional Binding
Repeated experiments have shown that voluntary
human actions are associated with systematic changes
in our perception of time [
            <xref ref-type="bibr" rid="ref4">3</xref>
            ]. The interval between a
voluntary or intentional action and the outcome of such
an actions is typically perceived as shorter than the
actual interval. For example, if a person voluntarily
presses a button and this action causes an outcome
e.g. a beep - it is highly likely that the person will
perceive their action as having happened later than
they it actually did (action binding). They are also likely
to perceive the outcome as having happened earlier
than it actually did (outcome binding). Patrick Haggard,
the research who first identified this phenomenon,
coined the term ‘Intentional Binding’ to describe it, as it
is contingent on several factors [
            <xref ref-type="bibr" rid="ref4">3</xref>
            ]. In the absence of
outcomes people are found to more accurately report
the timing of actions. For the temporal binding effect to
occur, actions must be intentional and must lead to an
outcome. Under these conditions our perception of the
timings of actions and their outcomes become bound
together temporally.
          </p>
          <p>In the years since Haggard’s first experiments, a large
number of studies have validated and built on his initial
observations. In and of itself this repeated
experimentation highlights the importance of replication
in cognitive neuroscience research. Based on this
replication, a scientific consensus is now supports the
conclusion that time perception in voluntary actions
and the binding effects associated with such actions
provides a robust implicit metric for the sense of
agency. Higher intentional binding values correlate to a
greater sense of personal agency.</p>
          <p>
            Replication and application
Detailed descriptions of the experimental methods used
to assess intentional binding are beyond the scope of
this short paper. These details are available in the CHI
2012 paper [
            <xref ref-type="bibr" rid="ref1">1</xref>
            ]. Instead I will focus more broadly on
the ways in which we replicated prior experiments and
applied this metric.
          </p>
          <p>Experiment 1
Neuro-cognitive experiments on intention binding
typically focus on very simple interactions, e.g. a
button press that causes a beep. My first experiment
focused on the modality of the interaction. It asked if
changes in the modality of an interaction lead to
changes in the sense of agency. The experimental
design closely mirrored procedures originally outlined
by Haggard. One independent variable was
manipulated: the input modality. We compared a
traditional input device - a keypad - with a skin-based
input device. The keypad replicated the input typically
used in neuroscience research. In condition one the
participant pressed a button on a keypad to cause a
beep. In condition two - the skin-based condition - the
participant caused a beep by tapping on their arm. The
skin-based capture device was attached to the
participant’s left arm and they tapped this arm with
their right hand. In all cases there was a fixed interval
of 250ms between the participant’s action and the
beep.</p>
          <p>Results showed that users experienced significantly
higher intentional binding for skin-based interactions.</p>
          <p>Across 19 participants a mean binding of 42.92ms was
observed in the button press condition. I.e. an interval
of 250ms was perceived as 207.08ms. Importantly, this
binding value is consistent with the results of prior
binding experiments that have used button inputs. In
the skin-based condition participants experienced a
total binding effect of 109.47ms. Here 250ms was
perceived as 140.53ms. Given the correlation between
intentional binding and the sense of agency, this
experiment suggests that people experience a
significantly greater sense of agency, or control, when
they interact with technology via skin-based input, as
compared with traditional keypad input.</p>
          <p>More broadly speaking, this experiment provided
empirical evidence that different interaction modalities
can provide different experiences of control and
ownership. In undertaking this experiment I believe it
was essential that one of our input conditions - the
keypad - replicated prior cognitive neuroscience
research. This replication demonstrates that our
experiment was administered effectively and lends
strength and credibility to our findings. It also allows
our results to be judged against and incorporated into
the prior body neuro-cognitive research on intentional
binding and the sense of agency.</p>
          <p>Ultimately I hope the method we introduced can be
used to investigate the sense of agency across a wide
range of input modalities. For other researchers using
this technique, I strongly recommend that replication
(plus extension) of prior results again be a key element
in the design of new experiments.</p>
          <p>Experiment 2
Cognitive neuroscience experiments on intentional
binding have typically examined voluntary and
involuntary actions. From an HCI perspective, this
might be considered an unnecessarily black or white
disjunction. Many user interactions with technology are
more intermediate. In particular ‘intelligent’ user
interfaces often seek to interpret and act on the
intentions of the user. Here users’ actions are
voluntary, but the outcomes may be assisted. The
second experiment in the CHI 2012 paper was designed
to investigate users’ sense of agency in interactions
where a computer interprets their intention and helps
them to achieve a goal. In this sense the second
experiment diverged further for the interactions
examined in prior cognitive neuroscience research.</p>
          <p>However, as in first experiment, we apply an
experimental procedure that closely matched prior
literature.</p>
          <p>The experiment investigated agency in a
machineassisted point-and-click task. Using a mouse,
participants were required to hit targets on a computer
screen, as quickly and as accurately as possible. The
computer provided assistance through an algorithm
that effectively added gravity to targets, thereby
making it easier for participants to complete the task.</p>
          <p>Hitting a target caused a beep. In each trial there was a
random interval between hitting a target and the beep,
and participants were asked to estimate this interval.</p>
          <p>In the experiment we investigated four different
assistance levels, which varied from no assistance to a
very high, and very obvious, level of computer
assistance. Results suggested that, up to a certain
point, the computer could assist users whilst also
allowing them to retain a sense of agency for their
actions. However, we found that beyond a certain level
of assistance users experienced a detectable loss in
their sense of agency. This loss in agency occurred in
spite of the fact that the computer correctly interpreted
users’ intentions and assisted them in achieving their
goal.</p>
          <p>Our results suggest that for the assisted input
algorithm we investigated - and possibly for assisted
input systems more generally - there may exist a
tipping point or sweet spot. This is the point at which a
computer can help people and potentially maximise
task performance - e.g. speed or accuracy - without
significant detriment to the experience of agency. I find
this possibility very intriguing. However I also believe
further investigation, and further replication, is required
to assess the generalizability of our initial finding. I am
currently undertaking such research.</p>
          <p>Conclusions
Alongside the issues discussed above, I have one minor
comment on the CHI submission process. When I
submitted the original CHI 2012 paper, I was very keen
to also submit the dataset for my studies. Under the
2012 submission system this was not possible. I
understand that this issue was addressed for CHI 2013
submissions. This was a real step forward.</p>
          <p>
            Citations
[
            <xref ref-type="bibr" rid="ref1">1</xref>
            ] Coyle, D., Moore, J., Kristensson, P.O., Fletcher,
P.C., &amp; Blackwell, A.F. (2012) I did that! Measuring
Users’ Experience of Agency in their own Actions. ACM
CHI 2012, 2025-2034.
[
            <xref ref-type="bibr" rid="ref2 ref3">2</xref>
            ] Ebert, J.P. &amp; Wegner, D.M., Time Warp: authorship
shapes the perceived timing of actions and events.
          </p>
          <p>
            Consciousness and Cognition, 2010. 19 481-89.
[
            <xref ref-type="bibr" rid="ref4">3</xref>
            ] Haggard, P., Clark, S., &amp; Kalogeras, J., Voluntary
action and conscious awareness. Nature Neuroscience
2002. 5(4).
          </p>
          <p>Replicating Two TelePresence Camera
Depth-of-Field Settings in One User
Experience Study
Jennifer Lee Carlson
Sr. User Experience Researcher
Cisco Systems, Inc.
170 W Tasman Drive
San Jose, CA 95134 USA
jennicar@cisco.com
Mike Paget
Sr. Technical Marketing Manager
Cisco Systems, Inc.
170 W Tasman Drive
San Jose, CA 95134 USA
mpaget@cisco.com
Tim McCollum
Sr. Design Manager
Cisco Systems, Inc.
170 W Tasman Drive
San Jose, CA 95134 USA
tmccollu@cisco.com</p>
          <p>Abstract
This paper describes an
experience study to
understand the user
perceptions on two
camera focus settings
in a TelePresence
room: limited- and
infinite-Depth-of-Field.</p>
          <p>The results influence
future TelePresence
experience design.</p>
          <p>Author Keywords
User experience;
comparative study;
TelePresence;
Depthof-Field; video
conferencing codec;
macroblocks; network
bandwidth</p>
          <p>ACM Classification Keywords
H.4.3. Information Systems: INFORMATION SYSTEMS
APPLICATIONS: Communications Applications
General Terms
Human Factors, Experimentation, Design
Introduction
Depth-of-Field is a description of the focal
characteristics within a captured image. It describes the
sharpness of the image from the foremost to farthest
areas on the z-axis within the cameras field of view.</p>
          <p>The cameras Depth-of-Field is determined by four key
factors which were related to works in this study:
1. The proximity of the two lenses to the camera
sensor and the cameras overall proximity to the
subject, otherwise known as focal length.
2. The amount of light that is allowed to reach the
sensor controlled by the aperture setting.
3. The duration at which light is allowed to pass
through the aperture, which is called shutter speed.
4. Camera gain setting, which can increase perceived
brightness in the image.
Two very common approaches to image capture
produce very different resulting images under the same
environmental conditions. The approach of limited
Depth-of-Field is to limit the amount of focal area
within the image to achieve controlled focal points.</p>
          <p>
            Generally this is an artistic decision for a particular
aesthetic style. This has also been used in video
applications such as cinema for the same artistic
purpose. However, in video applications such as
conferencing, the same image characteristics have
been used for a completely different purpose [
            <xref ref-type="bibr" rid="ref1">1</xref>
            ].
          </p>
          <p>Current video codecs used in conferencing systems
apply an algorithm that defines what information is
sent based on changed events rather than sending the
entire image. The algorithm groups areas of
information together in macroblocks and sends these
chunked updates when an area of the macroblock has
changed. This approach requires tedious preparation of
the environmental conditions and the camera settings.</p>
          <p>In the case of the Cisco TelePresence System
3000series (CTS-3xxx) system designs, they were purposely
built for a dedicated room that was optimized for very
high quality at a low network bandwidth. Therefore
they followed the model of camera settings that
provided limited Depth-of-Field.</p>
          <p>On the other hand, the approach of infinite
Depth-ofField, where a controlled focal point is not established,
is also common in both still image and video image
capture. This approach captures more detail within the
resulting image and requires the viewer (or end user)
to determine their own focal points as they view and
process the image. In a conferencing system this
approach will capture objects within the camera’s field
of view, as they exist without the need to adjust the
amount of sharpness. Such a system could require
additional processing power and bandwidth
requirements but it doesn’t require the same tedious
attention to detail of the environmental conditions or
the camera settings. Therefore this approach offers a
more flexible deployment model for a wider range of
conditions. In the case of the systems that utilized the
Precision HD camera, such as the Cisco TelePresence
3series (T3) system, they shared the same camera for
both dedicated and multipurpose room systems.</p>
          <p>Therefore use of an infinite Depth-of-Field configuration
was preferable to allow greatest amount of flexibility.</p>
          <p>The two depth-of-field applications were largely based
on technical and business reasons. What are the user
experience impacts, if any, from the two camera
settings in a TelePresence room? Our usability study
was to answer the following questions:
1. Are users aware of the difference in the two
camera focus approaches? If so, how do they differ?
2. Which approach feels more life-like to users? What
made it more life-like?
3. Which approach do users prefer and why? Are
there other considerations besides being life-like?
Methodology
In August 2011 the Cisco TelePresence User Experience
team conducted a formal usability study in an
immersive TelePresence room (see Figure 1).</p>
          <p>
            The study replicated the two camera settings in the
same TelePresence room to evaluate the user
experience in the context of a meeting. During the
session, users focused on the moderator, no documents
were shared, and the room had sufficient depth and
background to identify the moderator’s unique location.
We conducted a total of 27 within-subject comparative
usability study [
            <xref ref-type="bibr" rid="ref2 ref3">2</xref>
            ] sessions, with each session lasting
approximately 15 minutes. All participants have
experience with TelePresence.
Participants entered the TelePresence room containing
three side-by-side HD screens. The middle screen was
turned off during the entire study. The participant was
seated in the middle of the room so that the left and
right screens were the same distance from their seat.
          </p>
          <p>The left screen displayed an infinite Depth-of-Field,
where both moderator and background were in focus.</p>
          <p>The right screen displayed a limited Depth-of-Field,
where the moderator was in focus but the background
was blurred at a noticeable level. After the second day
of sessions (completed 15 participants) the background
objects were switched completely to counter-balance
any effects due to the background objects.</p>
          <p>In this study, in order to evaluate the user experience
impact from the two depth-of-field settings, it was
critical to make the other aspects of the images as
similar as possible, such as the field of view and the
subject matter within the frame. The two cameras
(used for the CTS-3xxx and T3 systems) for which we
wanted to test had very different physical
characteristics. But it was important that users couldn’t
tell the different cameras by their physical appearances
from the room. It was not possible to house both types
of cameras within the same system. Therefore, one
camera type was selected that fit the appropriate
physical characteristics as well as possessed settings
that could achieve both a limited depth-of-field and
infinite depth-of-field.</p>
          <p>Replicating the Depth-of-Field technique is relatively
easy in some areas and difficult in others. The lighting
and camera settings (hard and soft) can be replicated
easily with this controlled environment. The actual focal
settings are more challenging because we didn't
actually measure the depth of field with any equipment.</p>
          <p>It was assumed based on camera and light settings,
and by looking at the two set-ups subjectively.</p>
          <p>However, if we were to focus more effort on measuring
the depth of field as to define the distance and amount
of sharpness or blurriness, it could be more easily
replicated. The other area that was challenging to
replicate are the objects in the background. We setup
similar background based on props we had available.</p>
          <p>We could define more parameters on those props to
better replicate the testing.</p>
          <p>Procedures
Participants were told they would have a conversation
with a moderator via TelePresence to discuss their
experience with TelePresence, provide feedback and
rate their experience. Participants were not informed of
the difference in camera approaches until after they
had separately provided feedback and rated both
views. They looked at one view at a time until the very
end of the session when they compared the views side
by side.
Camera Setting 1
Participants were first presented with a view (segment)
of the moderator on either the right or left screen (the
order was reversed for every other participant to avoid
potential order effects). After answering
TelePresencerelated questions for several minutes, participants were
asked to rate the TelePresence session in terms of
video quality and how lifelike it appeared.</p>
          <p>Camera Setting 2
Then the view was switched to the opposite side of the
room and the moderator moved to the displayed view
to interact with the participant. After several minutes
of additional conversation, the participants were again
asked to rate the video quality and lifelike appearance
of the view.</p>
          <p>Comparisons
Participants were asked if they could tell any
differences between the two views they just looked at.</p>
          <p>If there were any differences, how the two views
appeared differently. Then they were shown both views
- one at a time - and asked if they noticed any
difference, or if they have noticed any other
differences. At the end, participants were shown both
views simultaneously so that they could make direct
comparisons. Participants were asked to describe any
differences they observed. If the participant could not
discern a difference in background clarity, the
moderator explained the differences between the
infinite and limited Depth-of-Field camera approaches.</p>
          <p>With this knowledge, participants then rated how
appealing each view was, which view they preferred
and why.</p>
          <p>A 7-point scale rating scale was used for all rating
questions, where 1 represented the ‘worst’ rating and 7
represented the ‘best’ rating.</p>
          <p>Findings
The study has identified the following key findings
based on participant behavior, feedback and preference
ratings:
1. Approximately 93% (25 of 27) participants were
unable to distinguish the camera focus approaches on
their own without viewing the images side by side.</p>
          <p>Even after viewing the images side by side, only 37%
(10 of 27) of participants were able to discern the
difference in background clarity between the two views.
2. Between the two camera focus approaches, on
average there were very minimal perceived differences
in terms of being lifelike (5.93 for infinite
Depth-ofField vs. 5.86 for limited Depth-of-Field) and video
quality (6.32 for infinite Depth-of-Field vs. 6.29 for
limited Depth-of-Field.)
3. After understanding the camera focus difference:
More participants (11 of 27 or 40%) preferred the
infinite Depth-of-Field approach. Fewer participants (8
of 27 or 30%) preferred the limited Depth-of-Field
approach. Almost one-third (8 of 27 or 30%)
participants did not have a preference between the two
approaches. On average the infinite Depth-of-Field view
was rated slightly more appealing (5.93 for infinite
Depth-of-Field vs. 5.52 for limited Depth-of-Field).</p>
          <p>Potential Future Work
This study was meant to be the first of a series of
studies. We want to find out what degree of camera
focus difference will be perceivable by most users. We
also want to study and analyze how user’s preferences
for camera focus relate to the different types of
meetings: such as an interactive brainstorming session,
a round-table team meeting, a single-speaker
presentation, or other types of meetings.
Acknowledgements
The authors thank all participants in Cisco Systems who
participated in the study described here. We also thank
Laura Borns of Cambridge Consultants for her
notetaking and analysis assistance for the study; Kevin
Nguyen and Rick AtKisson of Cisco Systems for their
support in TelePresence room set-up for the study, and
Chris Dunn of Cisco Systems, who initiated this
research study and reviewed this submission.</p>
          <p>
            References
[
            <xref ref-type="bibr" rid="ref1">1</xref>
            ] O'hara, K., Kjeldskov, J., Paay, J., Blended
interaction spaces for distributed team collaboration. In
ACM Transactions on Computer-Human Interaction
(TOCHI) TOCHI Homepage archive, Volume 18 Issue 1,
April 2011, Article No. 3.
[
            <xref ref-type="bibr" rid="ref2 ref3">2</xref>
            ] Sauro, J., Lewis, J.R. Quantifying the User
Experience: Practical Statistics for User Research
(2012, ISBN-10: 0123849683 | ISBN-13:
9780123849687), 10-11.
          </p>
        </sec>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>Punch</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          (
          <year>2002</year>
          ). Research With Children:
          <source>The Same or Different From Research With Adults? Childhood</source>
          <volume>9</volume>
          (
          <issue>3</issue>
          ),
          <fpage>321</fpage>
          -
          <lpage>341</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <surname>Morrow</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Richards</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          (
          <year>1996</year>
          ).
          <article-title>The Ethics of Social Research with Children: An Overview</article-title>
          .
          <source>Children &amp; Society</source>
          ,
          <volume>10</volume>
          (
          <issue>2</issue>
          ),
          <fpage>90</fpage>
          -
          <lpage>105</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [2]
          <string-name>
            <surname>Harper</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <year>2002</year>
          .
          <article-title>Talking about pictures: A case for photo elicitation</article-title>
          .
          <source>Visual Studies</source>
          ,
          <volume>17</volume>
          (
          <issue>1</issue>
          ),
          <fpage>13</fpage>
          -
          <lpage>26</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [3]
          <string-name>
            <surname>Pierce</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Schiano</surname>
            ,
            <given-names>D.J.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Paulos</surname>
          </string-name>
          , E. Home, habits, and
          <article-title>energy: examining domestic interactions and energy consumption</article-title>
          .
          <source>CHI</source>
          <year>2010</year>
          ,
          <year>1985</year>
          -
          <fpage>1994</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [4]
          <string-name>
            <surname>Rao</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sant</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Rajan</surname>
            ,
            <given-names>S.C.</given-names>
          </string-name>
          <article-title>An overview of Indian Energy Trends</article-title>
          .
          <year>2009</year>
          . Prayas, Energy Group, Pune, India.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [5]
          <string-name>
            <surname>Shrinivasan</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jain</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Seetharam</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Choudhary</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Huange</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dillahunt</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mankoff</surname>
            ,
            <given-names>J. CHI</given-names>
          </string-name>
          <year>2013</year>
          , (to appear).
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [6]
          <string-name>
            <surname>Thornberg</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <article-title>Informed grounded theory</article-title>
          .
          <source>Scandinavian Journal of Educational Research</source>
          ,
          <volume>56</volume>
          ,
          <year>2012</year>
          ,
          <fpage>243</fpage>
          -
          <lpage>259</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [7]
          <string-name>
            <surname>Vyas</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <article-title>Domestic Artefacts: Sustainability in the context of Indian Middle Class</article-title>
          .
          <source>ICIC</source>
          <year>2012</year>
          ,
          <volume>119</volume>
          -
          <fpage>128</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [8]
          <string-name>
            <surname>Woodruff</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hasbrouck</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <surname>Augustin. S.</surname>
          </string-name>
          <article-title>A bright green perspective on sustainable choices</article-title>
          .
          <source>CHI</source>
          <year>2008</year>
          ,
          <volume>313</volume>
          -
          <fpage>322</fpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>