<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Human Intelligence in Search and Retrieval</article-title>
      </title-group>
      <contrib-group>
        <aff id="aff0">
          <label>0</label>
          <institution>Carsten Eickho Department of Computer Science ETH Zurich</institution>
          ,
          <country country="CH">Switzerland</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>[3] Sebastian Deterding, Dan Dixon, Rilla Khaled, and Lennart Nacke. From game design elements to gamefulness: de ning gami cation. In Proceedings of the International Academic MindTrek Conference: Envisioning Future Media Environments. ACM</institution>
          ,
          <addr-line>2011</addr-line>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2015</year>
      </pub-date>
      <abstract>
        <p>[2] Ruggiero Cavallo and Shaili Jain. E cient crowdsourcing contests. In Proceedings of the International Conference on Autonomous Agents and Multiagent Systems. International Foundation for Autonomous Agents and Multiagent Systems, 2012. [4] Pei-Yun Hsueh, Prem Melville, and Vikas Sindhwani. Data quality from crowdsourcing: a study of annotation selection criteria. In Proceedings of the NAACL HLT 2009 workshop on active learning for natural language processing. Association for Computational Linguistics, 2009.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>Crowdsourcing has developed to become a magic
bullet for the data and annotation needs of modern
day IR researchers. The number of academic studies as
well as industrial applications that employ the crowd
for creating, curating, annotating or aggregating
documents is growing steadily. Aside from the multitude
of scienti c papers relying on crowd labour for system
evaluation, there has been a strong interdisciplinary
line of work dedicated to nding e ective and e cient
forms of using this emerging labour market. Central
research questions include (1) Estimating and
optimizing the reliability and accuracy of often untrained
workers in comparison with highly trained
professionals [1]; (2) How to identify or prevent noise and spam
in the submissions [4]; and (3) How to most
coste ciently distribute tasks and remunerations across
workers [2]. The vast majority of studies understands
crowdsourcing as the act of making micro payments
to individuals in return for compartmentalized units
of creative or intelligent labour.</p>
      <p>Gami cation proposes an alternative incentive
model in which entertainment replaces money as the
motivating force drawing the workers [3]. Under this
alternative paradigm, tasks are embedded in game
environments in order to increase the attractiveness and
immersion of the work interface. While gami cation
rightfully points out that paid crowdsourcing is not
the only viable option for harnessing crowd labour,
it is still merely another concrete instantiation of the
community's actual need: A formal worker incentive
model for crowdsourcing. Only by understanding
individual motivations can we deliver truly adequate
reward schemes that ensure faithful contributions and
long-term worker engagement. It is unreasonable to
assume that the binary money vs. entertainment
deCopyright c 2015 for the individual papers by the paper's
authors. Copying permitted for private and academic purposes.
This volume is published and copyrighted by its editors.
cision re ects the full complexity of the worker
motivation spectrum. What about education, socializing,
vanity, or charity? All of these are valid examples of
factors that compel people to lend us their work force.
This is not to say that we necessarily have to
promote edu cation and all its possible siblings as new
paradigms, they should merely start to take their well
deserved space on our mental map of crowdsourcing
incentives.</p>
      <p>In this talk, we will cover a range of interesting
scenarios in which di erent incentive models may
fundamentally change the way in which we can tap the
considerable potential of crowd labour. We will discuss
cases in which standard crowdsourcing and gami
cation schemes reach the limits of their capabilities,
forcing us to rely on alternative strategies. Finally, we will
investigate whether crowdsourcing indeed even has to
be an active occupation or whether it can happen as
a by-product of more organic human behaviour.
[1] Omar Alonso and Stefano Mizzaro. Can we get rid of
trec assessors? using mechanical turk for relevance
assessment. In Proceedings of the SIGIR 2009 Workshop
on the Future of IR Evaluation.</p>
    </sec>
  </body>
  <back>
    <ref-list />
  </back>
</article>