<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Cost and Quality Trade-Offs in Crowdsourcing</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Anja Gruenheid Systems Group ETH Zurich</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Donald Kossmann Systems Group ETH Zurich</string-name>
        </contrib>
      </contrib-group>
      <pub-date>
        <year>2013</year>
      </pub-date>
      <fpage>43</fpage>
      <lpage>46</lpage>
      <abstract>
        <p>Algorithms for crowdsourced tasks such as entity resolution, sorting, etc. have been subject to a variety of research work. So far, all of this work has focused on one specific problem respectively. In this paper, we want to focus on the bigger picture. More specifically, we want to show how it is possible to estimate the budget or the quality of an algorithm in a crowdsourcing environment where noise is introduced through incorrect answers by crowd workers. Such estimates are complex as noise in the information set changes the behavior of established algorithms. Using two sorting algorithms, QuickSort and BubbleSort as examples, we will illustrate how algorithms handle noise, which measures can be taken to make them more robust, and how these changes to the algorithms modify the budget and quality estimates of the respective algorithm. Finally, we will present an initial idea of how such an estimation framework may look like.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>INTRODUCTION</title>
      <p>
        Quality and its influence on the output result of any algorithm
using crowdsourcing has been subject to a range of research work.
Research focused on topics such as increasing the worker’s
motivation, quality control mechanisms, or accepting that workers make
mistakes and constructing fault-tolerant variations of the original
algorithms [
        <xref ref-type="bibr" rid="ref15 ref4">4, 15</xref>
        ]. This work will focus on none of these quality
assurance methods specifically but on assessing the inter-relationships
of the crowdsourcing budget, the crowd worker error rate, and
the result quality. Our goal is to define an intuition on how these
parameters are interleaved, meaning a) how for example adding
votes changes the result quality or b) if a certain result quality is
required, how many votes are required to meet these quality
constraints. These two scenarios are visualized in Figure 1. Obviously,
functions fQ and fB are closely related for the sake of conformity,
where fQ(Bi, perr) = Qi and fB (Qi, perr) = Bi holds for a
specific input budget Bi and input quality Qi. In other words, fQ is
the reverse calculation of fB . Note that the crowd worker error rate
is a given parameter in both scenarios that obviously influences the
result quality and budget requirements.
      </p>
      <p>It is not the objective of this work to define how this error rate can
Copyright c 2013 for the individual papers by the papers’ authors. Copying
permitted for private and academic purposes. This volume is published and
copyrighted by its editors.
algorithm B
perr
algorithm Q
perr</p>
      <sec id="sec-1-1">
        <title>Function fQ</title>
        <p>Q</p>
      </sec>
      <sec id="sec-1-2">
        <title>Function fB</title>
        <p>B
(a) Quality based on Budget</p>
        <p>
          (b) Budget based on Quality
be determined but we assume that it is a known parameter which
can be estimated through worker quality tests or more advanced
quality techniques [
          <xref ref-type="bibr" rid="ref5 ref9">5, 9</xref>
          ]. In addition to the crowd worker error rate
and the monetary budget (resp. the required quality), the estimation
function takes as input an algorithm which interprets the feedback
of the crowd. Depending on this algorithm and its efficiency in a
noisy environment, the estimated budget or quality may vary. For
the purpose of this paper, we will assume that the algorithm that we
want to estimate is a sorting algorithm but the quality and budget
estimation techniques that we will present here can also be used
for other algorithms in which the crowd is employed, such as entity
resolution.
        </p>
        <p>
          Independent of the algorithm, we assume that the quality of the
result decreases if the noise in the crowd answers increases which is
depicted in Figure 2. We will observe this behavior when examining
two example sorting algorithms, QuickSort and BubbleSort [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ].
For these algorithms, we will discuss quality and budget trade-offs,
showing how fault-tolerance mechanisms and specific adaptations to
the noisy environment can improve result quality effectively.
Additionally, we will show that the result quality and budget requirements
also highly depend on the applied algorithm and we present
examples where algorithms that were thought to be better alternatives
will produce results with lower quality.
        </p>
        <p>More specifically, we will first describe, using QuickSort and
BubbleSort as example algorithms, how different algorithms handle
noise with and without additional fault-tolerance mechanisms and
how they compare in terms of the amount of the budget that they
use and the quality of their results. From this specific example, we
will then draft the challenges of any estimation framework in this
kind of environment.</p>
      </sec>
    </sec>
    <sec id="sec-2">
      <title>RELATED WORK</title>
      <p>
        Crowdsourcing in context of database systems has been subject
to a variety of research recently. One research area focuses on
integrating crowdsourcing functionality efficiently into database
management systems while another research focus are algorithms
suitable for this kind of environment. Systems such as CrowdDB [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ],
Qurk [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], and Deco [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] connect traditional data storage systems
with crowdsourcing platforms such as Amazon Mechanical Turk
to gain additional information on queries [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. To enable users to
have similar functionality in those systems as in any comparable
relational database management system, research has further focused
on optimizing crowd accesses (i.e. reducing the budget spent on the
crowd) when implementing algorithms such as ranking and entity
resolution. For these algorithms, research has pursued two different
directions: The first class of algorithms focused on reducing the
budget while assuming that the crowd answers perfectly [
        <xref ref-type="bibr" rid="ref13 ref14">13, 14</xref>
        ],
observing the quality of the answers by the crowd workers as an
orthogonal problem. The second class of algorithms introduces the
notion of fault-tolerance to be able to handle noisy answers from
the crowd [
        <xref ref-type="bibr" rid="ref2 ref3 ref6 ref8">2, 3, 8, 6</xref>
        ].
      </p>
      <p>The novelty of our work is that we abstract from the actual
implementation of the algorithms in that we want to describe a more
general framework for the interdependencies between quality and
budget in a crowdsourcing environment independent of the applied
algorithm. Hence, we observe rather than construct the behavior of
a set of algorithms and show through examples which behavior fits
better for noisy environments.</p>
    </sec>
    <sec id="sec-3">
      <title>IMPACT OF NOISE</title>
      <p>Traditionally, sorting algorithms such as QuickSort and
BubbleSort have been evaluated with respect to time and space
requirements. The crowdsourcing context now adds another dimension,
namely result quality, as a noisy environment influences the quality
and correctness of the output of these algorithms negatively.
Moreover, temporal constraints become neglible as crowdsourcing itself
does not necessarily need immediate response algorithms by
design while space requirements can be seen as budget constraints
(i.e. the number of comparisons that are required to find a solution).
We will show in the following that the choice of algorithm for a
certain problem heavily influences the result quality especially in
noisy environments. To that purpose, we will first evaluate
QuickSort in a crowdsourcing environment after which we will focus on
BubbleSort as example algorithms.
3.1</p>
    </sec>
    <sec id="sec-4">
      <title>QuickSort</title>
      <p>The prerogative of QuickSort is that it can efficiently sort an input
set, using O(n log(n)) comparisons on average. Thus, it is a prime
candidate for sorting in a crowdsourcing environment as it allows
to reduce the budget spent on the task. On the other hand, noise in
the crowdsourcing answers directly propagates if QuickSort is used.
An example for this observation is shown in Figure 3. Imagine a
crowdsourcing task which requires the workers to order a few words
according to their positivity. In the first example answer set, none
of the edges (i.e. the votes of crowd workers) is wrong which leads
to low budget result with perfect quality. In contrast, in the second
1st pivot: good</p>
      <p>good
2nd pivot: bad
excellent
bad
1st pivot: good
2nd pivot:
bad, excellent
neutral
painful
(a) QuickSort Without Noise</p>
      <p>good
excellent
bad
painful</p>
      <p>neutral
(b) QuickSort With Noise
example set the edge between ‘painful’ and ‘good’ distorts the result
in that painful is now ranked higher than three other words due to
the transitive propagation of this specific wrong comparison.</p>
      <p>
        How noise affects the output of the QuickSort algorithm is
dominated by the distribution of the noisy answers and whether the
implementation of QuickSort used in this context is fault-tolerant.
Many of the current crowdsourcing solutions do not provide
faulttolerant algorithms but assume the crowd to answer perfectly [
        <xref ref-type="bibr" rid="ref13 ref14">13,
14</xref>
        ]. In contrast, we believe that given the application of these
algorithms where information is collected in crowdsourcing platforms
such as Amazon Mechanical Turk, accepting that the crowd does
create a certain amount of noise is only natural. Mechanisms that
can make QuickSort fault-tolerant may include strategies such as
using the majority of the crowd votes in order to determine an edge
or to request that one answer exceeds the alternative by quorum
votes. Depending on the strategy, the estimates of functions fQ and
fB vary. For example imagine that perr=.2, meaning here that 20%
of the actual comparisons are wrong on average. If the majority
strategy is applied and all false votes are countered by more true
votes, the quality of the result can still be impeccable. On the other
hand, if no fault-tolerant mechanism is used, the result quality will
suffer because wrong comparisons will be propagated directly into
the result. Obviously, this is the best case scenario for the majority
strategy and in practice even with fault-tolerance comparisons will
be propagated wrongly into the decision structure. But given the
integrated quality assurance mechanism of the majority strategy, a
smaller number of comparisons will be wrong which means higher
result quality in the end. Thus, our first observation is that depending
on the interpretation strategy used (for example the direct
propagation of votes or the majority strategy explained in the previous
example), the budget and the quality output changes.
      </p>
      <p>Figure 4 depicts a simulated budget-quality trade-off for the two
discussed strategies, the direct propagation of votes and the majority
interpretation strategy. It shows that fault-tolerance increases the
budget requirements for QuickSort but it also improves the result
quality, i.e. the number of pair-to-pair comparisons that are
confirmed by the ground truth. To better understand this observation
take again the running example: If the crowd worker error rate is
quality
1
.5
perr=0
perr=.1, direct
perr=.2, direct
perr=.1, majority
perr=.2, majority
31500
cost
perr=.2, at least one edge is faulty out of the six that we need for
our QuickSort solution in Figure 3. The result quality of our
example is improved if the majority interpretation strategy is applied
because on average, the erroneous votes are better distributed than
in the direct scenario thus resulting in less erroneous comparisons
as exemplified above. The figure also visualizes that applying the
majority strategy leads to a higher budget as a comparison is made
multiple times, thus the overall information gain is slower per unit
of budget.</p>
      <p>During the simulations a decrease in cost can be observed when
the error rate increases, i.e. the cost for QuickSort with the direct
strategy and without noise is approximately 11,000 while the cost
decreases for perr = .1 to 9,700. The explanation for this behavior
is that in cases where the pivot element is not a central element in
the final sorting, wrong answers may benefit the overall number of
comparisons as the wrongly judged item is transfered into a smaller
bucket and thus used less often overall for comparisons.</p>
      <p>Overall, the main observation for QuickSort here is that if the
error increases, fault-tolerance measurements can help to improve
the quality of the result set even if they also mean an increase in
budget in order to reach the break-even point. On the other hand, it
is essential to observe that even with fault-tolerant techniques, there
is no guarantee that QuickSort will provide a correct result to the
problem of sorting an input dataset. Thus, if the problem statement
is to find a complete ranking of the input set, QuickSort fails as a
sorting algorithm in this example setting.
3.2</p>
    </sec>
    <sec id="sec-5">
      <title>BubbleSort</title>
      <p>In contrast to QuickSort, BubbleSort is considered to be
suboptimal in terms of the number of comparisons it needs to find a solution
as it does local comparisons rather than global comparisons and its
average comparison complexity is O(n2). In a noisy environment
predicates shift and the number of pure item comparisons is not the
decisive factor anymore as we will show in the following.
BubbleSort specifically has one characteristic that makes this algorithm
suitable for noisy environments: It benefits from the input order of
the records. Thus, if the input is already sorted, the algorithm will
run in O(n). This characteristic can be leveraged by running
BubbleSort multiple times over the same input dataset. Due to the input
sensitivity of BubbleSort, subsequent runs will have a decreased
number of comparisons. BubbleSort in a noisy environment will
further benefit from a very intuitive modification to the algorithm that
is made due to the cost sensitivity of the environment: Every
comparison that is requested is stored. As a result, comparing the same
item twice will not result in additional cost but if a crowd worker
returns the wrong answer this answer will be propagated through the
dataset. Even though this kind of propagation reduces the quality
of the output, the cost of iteratively improving the sorting solution
is less than running BubbleSort without storage mechanisms as
quality
1
runs
verified through simulations.</p>
      <p>These modifications are minor and do not change the character
of the algorithm which proceeds by doing local comparisons and
‘bubbling’ items to their correct spot in the ordering.
3.3</p>
    </sec>
    <sec id="sec-6">
      <title>QuickSort vs. BubbleSort</title>
      <p>When comparing QuickSort and BubbleSort and observing their
behavior in a noisy setting, we can see that traditional algorithmic
assumptions shift. Consider Figure 5 which shows an example
sketch for this shift: Here, we draw the behavior of the direct
QuickSort approach and the BubbleSort variation described above
in terms of cost and quality simulated in a noisy environment with
perr = .2 and 1000 items to be sorted. As we want to obtain
the optimal result quality eventually, we execute both algorithms
multiple times, the number of runs in this simulation is 500. There
are two main observations that can be drawn from these simulations.</p>
      <p>First, they show that BubbleSort can improve its output result
over time due to its input sensitivity while the output quality of
QuickSort remains constant. Additionally, the average number of
crowd comparisons requested in BubbleSort decreases over time (in
our simulations up to 50%). Thus, the output quality of BubbleSort
increases because if the algorithm asks for fewer input from the
crowd, fewer noise is introduced into the sorting. In contrast, the
level of noise that the QuickSort algorithm has to handle remains the
same over multiple runs. The second observation that can be drawn
from these simulations is that even if cost reduction techniques are
used, BubbleSort is in this setup 26 times more expensive than
QuickSort when achieving the same quality (i.e. at the quality
breakeven point). Thus, even though the increase in cost is lower, the
necessity of repeating the algorithm multiple times increases the
overall cost for BubbleSort in the end.</p>
      <p>The example above shows that some algorithms like
BubbleSort have a natural robustness that makes them more suitable in
noisy environments. Thus, we think that established algorithms that
outperform others according to traditional complexity theory need
to be looked upon from a new angle in noisy environments as an
alternative or as a component in hybrid setups.
4.</p>
    </sec>
    <sec id="sec-7">
      <title>QUALITY &amp; BUDGET ESTIMATES</title>
      <p>In order to realize an estimation framework, we propose as first
step to establish a modified complexity theory that is noise-aware.
That is, in traditional complexity theory the number of comparisons
per algorithm determines the usefulness of that specific algorithm.
As we have shown previously, traditional assumptions may not hold
in the crowdsourcing environment due to noise in the information
set. Thus, algorithms need to be evaluated taking into consideration
their behavior in noisy environments. This evaluation can then be
used to determine which algorithm is most suitable to solve a certain
problem. As a result, we propose to define a new model for the
complexity of an algorithm that is dependent on a) the number of
records in the input set (analogous to traditional complexity theory)
b) and the amount of noise that this algorithm will encounter.</p>
      <p>If the complexity of an algorithm is known, we can then use it
to form estimates of the required budget given certain quality
constraints or estimates that approximate the result quality given certain
budget constraints for an algorithm as shown initially in Figure 1.
Note that this notion of complexity is not restricted to sorting
algorithms but it can be used for evaluating a variety of algorithms such
as entity resolution, finding the maximum etc. To be able to take
input parameters such as a required quality or available budget,
traditional algorithms need to be adjusted. Modifying these algorithms
is necessary because current algorithms take an input dataset after
which they are executed to return a result with a certain quality for
a certain budget independent of the crowd worker error rate. How
budget and quality are distributed is subject to the specific execution
run and resemble points on a curve similar to those depicted in
Figure 5 where the curve varies according to the level of noise in
the information set and the chosen algorithm.</p>
      <p>In our opinion, algorithms need to be conscious of input
constraints to be able to compare two algorithms in a noisy environment.
Going back to our comparison of QuickSort and BubbleSort, we
can determine that according to the characteristics of the algorithms,
QuickSort with fault-tolerance can result in good quality results in
low noise environments while BubbleSort can be more adequate in
environments where the crowd answers wrongly more often. But in
order to be able to define the actual trade-off, we need to compare
how the algorithms react in scenarios with fixed error rate and
budget or quality. Thus, we want to be able to define the cost-quality
function curves instead of points on the curve that can be derived
from current algorithm variations that are not constraint-aware as
mentioned previously.</p>
      <p>If these curves are known, we can use them for our estimation
framework in that they define the complexity of an algorithm
dependent on the crowd worker error rate where we can vary budget and
quality constraints in order to find a suitable estimated output. A
simple two-step estimation process may then look as follows:
1. Initial Estimate - Given the properties of the chosen
algorithm, i.e. its complexity in this kind of noisy environment,
decide upon an initial estimate.
2. Adjust Estimate - Using the knowledge of the input
quality respectively budget parameter, adjust the estimate as to
meet/exhaust the parameter. An example measure is to
allocate more budget for multiple runs of the same algorithm or
for fault-tolerant mechanisms.</p>
      <p>Obviously, this process is only a rough sketch of the functionality
of such an estimation framework. More specific implementation
and design details of these two steps are subject to future work as
is the specification of the complexity theory that enables us to give
accurate estimates in noisy environments.</p>
    </sec>
    <sec id="sec-8">
      <title>CONCLUSION</title>
      <p>In this work, we have presented several observations made for
quality and cost trade-offs in the still novel crowdsourcing
environment. Using QuickSort and BubbleSort as example algorithms, we
have shown that any estimation function that wants to determine the
expected output of that algorithm has to take into consideration that
different algorithms behave differently when confronted with noise.
Their behavior not only changes in terms of budget requirements but
they also vary in their robustness in noisy environments. We have
presented techniques such as fault-tolerant interpretation strategies
and increasing the number of executions of the algorithms that
influence the result estimation and choice of algorithm when taken into
consideration. Last, we have shown that traditional beliefs such that
QuickSort is better than BubbleSort due to its lower number of
comparisons are questioned in a noisy environment because even though
QuickSort is requires a smaller budget, BubbleSort may provide
better quality results due to its ingrained second chance mechanism
and its ability to leverage information over multiple runs.</p>
      <p>In the future, we plan to specify our sketched estimation
framework and to define exactly how noise can be evaluated in such an
environment. Additionally, we want to determine whether it is
possible to formally analyze the trade-off between cost and quality for
algorithms that are traditionally integrated into relational database
management systems. Knowing the complexity of these algorithms
will help us to determine which of them are viable options for a
database system that incorporates crowdsourcing as a way to obtain
information.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>M. J.</given-names>
            <surname>Franklin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Kossmann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Kraska</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Ramesh</surname>
          </string-name>
          , and
          <string-name>
            <given-names>R.</given-names>
            <surname>Xin</surname>
          </string-name>
          .
          <article-title>Crowddb: answering queries with crowdsourcing</article-title>
          .
          <source>In SIGMOD Conference</source>
          , pages
          <fpage>61</fpage>
          -
          <lpage>72</lpage>
          ,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>A.</given-names>
            <surname>Gruenheid</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Kossmann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Ramesh</surname>
          </string-name>
          , and
          <string-name>
            <given-names>F.</given-names>
            <surname>Widmer</surname>
          </string-name>
          .
          <article-title>Crowdsourcing entity resolution: When is a=b? Technical report</article-title>
          , ETH Zurich,
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>S.</given-names>
            <surname>Guo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Parameswaran</surname>
          </string-name>
          , and
          <string-name>
            <given-names>H.</given-names>
            <surname>Garcia-Molina</surname>
          </string-name>
          .
          <article-title>So Who Won? Dynamic Max Discovery with the Crowd</article-title>
          .
          <source>In Proceedings SIGMOD</source>
          ,
          <year>2012</year>
          . to appear.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>A.</given-names>
            <surname>Kittur</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. V.</given-names>
            <surname>Nickerson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Bernstein</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Gerber</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Shaw</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Zimmerman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Lease</surname>
          </string-name>
          , and
          <string-name>
            <surname>J. Horton.</surname>
          </string-name>
          <article-title>The future of crowd work</article-title>
          .
          <source>In Proceedings of the 2013 conference on Computer supported cooperative work</source>
          ,
          <source>CSCW '13</source>
          , pages
          <fpage>1301</fpage>
          -
          <lpage>1318</lpage>
          , New York, NY, USA,
          <year>2013</year>
          . ACM.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>M.</given-names>
            <surname>Lease</surname>
          </string-name>
          .
          <article-title>On quality control and machine learning in crowdsourcing</article-title>
          .
          <source>In Human Computation</source>
          ,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>A.</given-names>
            <surname>Marcus</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Karger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Madden</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Miller</surname>
          </string-name>
          , and
          <string-name>
            <given-names>S.</given-names>
            <surname>Oh</surname>
          </string-name>
          .
          <article-title>Counting with the crowd</article-title>
          .
          <source>In Proceedings of the 39th international conference on Very Large Data Bases, PVLDB'13</source>
          , pages
          <fpage>109</fpage>
          -
          <lpage>120</lpage>
          . VLDB Endowment,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>A.</given-names>
            <surname>Marcus</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. R.</given-names>
            <surname>Karger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Madden</surname>
          </string-name>
          , and
          <string-name>
            <given-names>R. C.</given-names>
            <surname>Miller</surname>
          </string-name>
          .
          <article-title>Demonstration of qurk: a query processor for humanoperators</article-title>
          .
          <source>In SIGMOD Conference</source>
          , pages
          <fpage>1315</fpage>
          -
          <lpage>1318</lpage>
          ,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>A.</given-names>
            <surname>Marcus</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. R.</given-names>
            <surname>Karger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Madden</surname>
          </string-name>
          , and
          <string-name>
            <given-names>R. C.</given-names>
            <surname>Miller</surname>
          </string-name>
          .
          <article-title>Human-powered sorts and joins</article-title>
          .
          <source>PVLDB</source>
          ,
          <volume>5</volume>
          (
          <issue>1</issue>
          ):
          <fpage>13</fpage>
          -
          <lpage>24</lpage>
          ,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>D.</given-names>
            <surname>Oleson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Sorokin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G. P.</given-names>
            <surname>Laughlin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Hester</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Le</surname>
          </string-name>
          , and
          <string-name>
            <given-names>L.</given-names>
            <surname>Biewald</surname>
          </string-name>
          . Programmatic gold:
          <article-title>Targeted and scalable quality assurance in crowdsourcing</article-title>
          .
          <source>In Human Computation</source>
          ,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>A.</given-names>
            <surname>Parameswaran</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Park</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Garcia-Molina</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Polyzotis</surname>
          </string-name>
          , and
          <string-name>
            <given-names>J.</given-names>
            <surname>Widom</surname>
          </string-name>
          . Deco:
          <article-title>Declarative crowdsourcing</article-title>
          .
          <source>Infolab Technical Report</source>
          , Stanford University,
          <year>November 2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>R.</given-names>
            <surname>Sedgewick</surname>
          </string-name>
          and
          <string-name>
            <given-names>K.</given-names>
            <surname>Wayne</surname>
          </string-name>
          . Algorithms, 4th Edition. Addison-Wesley,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>B.</given-names>
            <surname>Trushkowsky</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Kraska</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. J.</given-names>
            <surname>Franklin</surname>
          </string-name>
          , and
          <string-name>
            <given-names>P.</given-names>
            <surname>Sarkar</surname>
          </string-name>
          .
          <article-title>Crowdsourced enumeration queries</article-title>
          .
          <source>In ICDE</source>
          ,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>J.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Kraska</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. J.</given-names>
            <surname>Franklin</surname>
          </string-name>
          , and
          <string-name>
            <given-names>J.</given-names>
            <surname>Feng</surname>
          </string-name>
          . Crowder:
          <article-title>Crowdsourcing entity resolution</article-title>
          .
          <source>PVLDB</source>
          ,
          <volume>5</volume>
          (
          <issue>11</issue>
          ):
          <fpage>1483</fpage>
          -
          <lpage>1494</lpage>
          ,
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>S. E.</given-names>
            <surname>Whang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Lofgren</surname>
          </string-name>
          , and
          <string-name>
            <given-names>H.</given-names>
            <surname>Garcia-Molina</surname>
          </string-name>
          .
          <article-title>Question selection for crowd entity resolution</article-title>
          .
          <source>Technical report</source>
          , Stanford University.
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>X.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Fan</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Y.</given-names>
            <surname>Yu</surname>
          </string-name>
          . Sembler:
          <article-title>Ensembling crowd sequential labeling for improved quality</article-title>
          .
          <source>In AAAI</source>
          ,
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>