<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Means Better Data: a Crowdsourcing Lab</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Ryan J. A. Murphy</string-name>
          <email>rmurphy@mun.ca</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Jeffrey Parsons</string-name>
          <email>jeffreyp@mun.ca</email>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Memorial University of Newfoundland</institution>
          ,
          <addr-line>230 Elizabeth Avenue, St. John's, NL</addr-line>
          ,
          <country country="CA">Canada</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Data granularity is the level of direct correspondence between data in an Information System (IS) and the real-world things represented by the data. It determines the amount of detail that may be captured, stored, and used by contributors and consumers of information in an IS. We present a between-groups lab experiment in which we manipulated the granularity of a crowdsourcing project's interface to assess the impact of granularity on data completeness, data correctness, and overall contributor participation. We found that contributors using a finer-grained data collection interface contributed more complete data, while contributors using a coarser-grained data collection interface contributed more incorrect data. Moreover, the level of granularity did not influence the degree of participation. These findings suggest that granularity is an important issue in the design of data crowdsourcing projects.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Data crowdsourcing</kwd>
        <kwd>observational crowdsourcing</kwd>
        <kwd>granularity</kwd>
        <kwd>conceptual model</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        How does the design of data crowdsourcing projects influence the level of detail contributors are
able and willing to contribute? How does the level of detail contributors are able to contribute influence
their ability to contribute complete and correct data? In this paper, we explore how granularity (the level
of direct correspondence between data in an Information System (IS) and the real-world things
represented by that data [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]) influences the level of completeness and correctness in data captured by a
crowd in a data crowdsourcing project.
      </p>
      <p>
        Data crowdsourcing is a phenomenon in which a crowd is mobilized to collect or analyze large
volumes, varieties, and/or velocities of data [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]—of potentially-questionable veracity [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. Observational
crowdsourcing is a kind of data crowdsourcing in which contributors capture observations about some
domain (e.g., wildlife) of the real world over a continuous period [
        <xref ref-type="bibr" rid="ref4 ref5">4, 5</xref>
        ].
      </p>
      <p>
        An important opportunity for data crowdsourcing projects is the unanticipated use and reuse of
collected data [
        <xref ref-type="bibr" rid="ref6 ref7">6, 7</xref>
        ]. Yet, once data has been collected, it can be very difficult or even impossible to
return to the observation that was the object of that data and capture more detail from it. For this reason,
data granularity is an important issue for data crowdsourcing (and especially for observational
crowdsourcing). Granularity is an important factor in the ability to use and reuse data. If data is captured
at finer-grained levels of detail (e.g., features and descriptions of observed wildlife), it may be possible
to combine the collected details in useful ways. However, if data is collected at coarse-grained levels
(e.g., classes, such as type of animal), then potentially important details about the observation cannot
be captured and may be lost forever [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ].
      </p>
      <p>However, before we can leverage the granularity in the design of data crowdsourcing projects, we
need to guarantee that fine-grained data collected from data crowdsourcing actually does contain more
detail than coarse-grained data. Moreover, we must make sure that this detail is useful (e.g., does it
facilitate more complete representation of the real-world phenomena the observer is capturing?) and
that it is correct (e.g., finer-grained data capture does not introduce errors). So, in the present research,
we explore three related research questions: (1) Are observations captured at finer levels of granularity
more or less complete than observations captured at coarser levels of granularity? (2) Are observations
captured at finer levels of granularity more or less correct than those captured at coarser levels of
granularity? (i.e., does finer-grained data cause contributors to introduce more errors?) (3) As
contributing fine-grained data may require more effort than coarse-grained data, do contributors
providing fine-grained observations contribute fewer overall observations?
2. Experimenting with granularity in crowdsourced data collection</p>
      <p>One way to expose the conceptual model of an information system to data contributors is through
the interface used to submit data. The interfaces data contributors use shape the data they contribute, as
prompts, input boxes and so on shape and frame the structure, type, and content of contributions.
Therefore, in this experiment, we use different degrees of granulation in data collection interfaces to
represent different levels of granularity in the hypothetical project’s conceptual model.</p>
    </sec>
    <sec id="sec-2">
      <title>2.1. Hypotheses 2.1.1. Data completeness</title>
      <p>
        Wand and Wang [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] define completeness, an intrinsic dimension of data quality, as “the ability of
an information system to represent every meaningful state of the represented real world system” (p. 93).
Granularity clearly plays a role in enabling this ability in information systems: if the conceptual model
of the IS is insufficiently granulated to represent some details of the real world, the data captured within
that IS (determined by data collection and storage decisions) cannot include those details. An IS with a
more finely-granulated conceptual model will be more able to completely represent a given domain
than an IS with a less granulated conceptual model. At the same time, it may be unnecessary to
maximally granulate a conceptual model in order to benefit from the effects of granularity. A
coarsegrained model may facilitate a similar level of completeness as a fine-grained model. While both would
encourage contributors to break down their contributions into more detail, at a certain threshold that
level of detail may be tedious, leading to diminishing returns on increasing levels of granularity. We
are therefore interested in the degree of benefit fine granularity and coarse granularity each provide
over an ungranulated alternative. This gives us our first set of hypotheses:
• H(1a): An IS with a fine-grained data collection interface will generate more complete data
than an IS with an ungranulated data collection interface.
• H(1b): An IS with a coarse-grained conceptual model will generate more complete data than
an IS with an ungranulated data collection interface.
      </p>
      <p>Note that we do not hypothesize a difference between coarse-grained and fine-grained interfaces in
terms of the completeness of data collected.</p>
    </sec>
    <sec id="sec-3">
      <title>2.1.2. Data correctness</title>
      <p>
        We follow Wand and Wang’s [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] definition of data correctness: the degree to which an information
system’s data represents valid states of its real-world domain. The more finely granulated the conceptual
model of an IS is, the more specifically it can represent the target domain. Coarsely granulated
conceptual models create data that depend on user inference to fill in details. In turn, we propose that
less granulated conceptual models lead to data that is more prone to incorrectness. This gives us our
second set of hypotheses (Note that we do not hypothesize a difference between coarse-grained and
fine-grained interfaces in terms of the correctness of data collected):
• H(2a): An IS with a fine-grained data collection interface will generate less incorrect data
than an IS with an ungranulated data collection interface.
• H(2b): An IS with a coarse-grained data collection interface will generate less incorrect data
than an IS with an ungranulated data collection interface.
      </p>
    </sec>
    <sec id="sec-4">
      <title>2.1.3. Granularity and cognitive effort</title>
      <p>•</p>
      <p>As discussed above, we expect granulated data to be more complete and more correct than
ungranulated data. However, capturing this degree of completeness and correctness will be more
cognitively demanding on users. Consequently, we expect that contributors will make fewer
contributions when data collection is granulated than when it is ungranulated. This gives us two final
hypotheses:
•</p>
      <p>H(3a): Users contributing data to an information system with an ungranulated data
collection interface will produce more contributions than users contributing data to an
information system with a fine-grained data collection interface.</p>
      <p>H(3b): Users contributing data to an information system with an ungranulated data
collection interface will produce more contributions than users contributing data to an
information system with a coarse-grained data collection interface.
2.2.</p>
    </sec>
    <sec id="sec-5">
      <title>Methodology</title>
      <p>To explore how granularity influences data completeness, we ran a between-groups experiment in
which we manipulated the data collection interface of a data crowdsourcing system to instantiate three
different levels of conceptual model granularity. Participants were randomly assigned to one of the
three conditions. After reviewing instructions for the task, completing a simple comprehension test, and
responding to some background questions (e.g., self-report measures of wildlife expertise), participants
were presented a set of up to 20 photos of wildlife, one at a time, in random order. Their task was to
describe each photo using the system interface. Each participant completed a minimum of five such
observations, after which they could choose to end their participation at any time. After participants
opted to exit, or after they completed all 20 observations, they were asked a few questions about their
experience in the before concluding their participation in the experiment. The experimental
manipulation was never revealed to participants—i.e., participants in one condition were not told about
the existence of alternative interfaces. More details about the experiment’s participants, materials, and
resulting measures are described below.</p>
    </sec>
    <sec id="sec-6">
      <title>2.2.1. Participants</title>
      <p>One hundred participants were recruited via Amazon Mechanical Turk (MTurk), a crowdsourcing
platform in which people receive money for completing micro-tasks. Participants received $2 USD as
a reward for completely participating in the experiment. 41 participants were randomly assigned to the
“ungranulated” condition, 32 to the “coarse granularity” condition, and 27 to the “fine granularity”
condition. Several participants (six from the “ungranulated” condition, five from the “coarse
granularity” condition, and two from the “fine granularity” condition) seemed to complete each
observation by using the sample photo to complete a reverse-image search on Google, then copying
and pasting information from the first result into the task text boxes. These results were discarded,
leaving 87 participants: 35 in the “ungranulated” condition, 27 in the “coarse granularity” condition,
and 25 in the “fine granularity” condition.</p>
    </sec>
    <sec id="sec-7">
      <title>2.2.2. Materials</title>
      <p>In the MTurk interface, participants were given a brief overview of the experiment (essentially
explaining that they would be helping researchers describe photos of wildlife), an informed consent
form, a hyperlink to the task, and a text box in which to paste a completion code received after
successfully completing the task.</p>
      <p>The hyperlink in the MTurk interface brought participants to the experimental task materials,
developed using Qualtrics survey software. Upon arrival, participants were presented with the task
instructions. After reviewing the instructions, participants needed to successfully answer two questions
testing their basic comprehension of the task. They were then asked four questions about their
background before beginning the task itself. These materials are provided in Appendix 1.
2.2.3. Task</p>
      <p>After completing the background questions, participants began the experimental task. Each
participant completed between five and 20 observations in which they described the contents of a photo
of wildlife. We allowed participants to exit after a minimum of five completed observations in order to
assess whether our granularity manipulation influenced the number of contributions a participant was
willing to make. This design allowed us to evaluate the influence of the experimental conditions in a
setting without an extrinsic motivator. If participants completed fewer observations in any given
condition (e.g., if they dropped out before completing 20 observations), it would indicate that the
condition involved a more difficult or aversive task compared to a condition where participants
completed more observations. The sample photos were taken from NL Nature (www.nlnature.com), a
crowdsourcing platform in which users contribute sightings of wildlife in Newfoundland and Labrador.
Up to twenty different photos of wildlife were presented to each participant in random order. For each
observation, we measured the number of seconds between when the observation was first loaded and
when the participant submitted the observation to exit the task or to move on to the next observation.</p>
      <p>For each observation, depending on the condition to which they were assigned, participants were
presented with a screen like one of the three shown in Figure 2. While each participant had to complete
the same task—describing the wildlife in the photo—we modified the ways in which participants had
to provide their description across each condition. In the “Ungranulated” condition, participants were
asked “What do you see?”, seeking to emulate similar interfaces in real-world data crowdsourcing
interfaces. In the “Coarse granularity” condition, participants were instead asked “What kind of wildlife
do you see?”, “What are the features of the wildlife?”, and “What are the behaviours of the wildlife?”—
these three questions granulate the question “what do you see?” into three major possible groups of
observations. Last, in the “Fine granularity” condition, participants were asked “What kind of wildlife
do you see?”, “Colour(s) of the head:”, “Features of the head:”, “Colour(s) of the tail:”, “Features of
the tail:”, “Colour(s) of the limbs:”, “Features of the limbs:”, “Colour(s) of the body:”, “Features of the
body:”, “What behaviours do you see?”, and “What is it interacting with?”, further granulating the
questions in the coarse-grained category. To maintain structural equivalence, each condition included
only 11 text boxes of identical size, exactly as depicted in the figure. Participants were instructed only
to report what they could actually see, not to guess based on what kind of wildlife they thought they
were looking at. Participants were also instructed not to fill in text boxes if it wasn’t possible to answer
a given prompt.</p>
    </sec>
    <sec id="sec-8">
      <title>Results</title>
      <p>Analysis was completed with SPSS Statistics version 27. As seen in Table 2, participants in the
Ungranulated condition completed an average of 9.17 (std. dev. = .86) observations (e.g., they looked
at and described 9.17 photos of wildlife, on average). Participants in the Coarse Granularity condition
completed an average of 9 (std. dev. = .89) observations, and participants in the Fine Granularity
condition completed an average of 7.68 (std. dev. = .94) observations. The results presented here
therefore include analysis of 756 total observations from 87 participants across the three conditions.
n
35
27
25
n
35
27
25</p>
      <p>Using a one-way ANOVA (Table 2), we found no significant difference between groups’
selfreported biology/ecology education (F(2, 84) = .692, p = .882), identification as citizen scientists (F(2,
84) = .750, p = .475), or hours spent outdoors (F(2, 84) = .297, p = .744). However, we found that
participants’ self-report of wildlife expertise violated Levene’s test of homogeneity of variances, F(2,
84) = 6.576, p = .002. Therefore we used a one-way ANOVA with Welch’s adjusted F ratio to test for
significance, again finding none, F(2, 55.683) = .597, p = .554.</p>
      <p>To explore potential differences in task completion time (Table 3), we analyzed the difference
between the average completion time per-observation in each condition. These measures violated
Levene’s test for equality of variances, F(2,84) = 3.648, p = .030. So, we conducted a one-way ANOVA
with Welch’s adjusted F ratio, finding a statistically significant difference in mean observation
completion time per-participant between conditions, F(2, 49.918) = 3.619, p = .034. Because group
sizes were uneven and variances were unequal, we used Games-Howell’s post-hoc procedure for
multiple comparisons, finding only a significant mean difference between the fine-grained and
coarsegrained conditions (mean difference = 48.68, p (.035) &lt; α (.05).)</p>
      <p>Finally, we checked to see if participants in any condition had more or less confidence, found the
task easier or more difficult, or more or less enjoyable than the other conditions. To do this, we
conducted a one-way ANOVA on each measure, finding no statistically significant difference in
participant confidence (F(2, 84) = .177, p = .838), ease (F(2, 84) = .128, p = .838), or enjoyment (F(2,
84) = 1.609, p = .206) across the three conditions (Table 3).</p>
    </sec>
    <sec id="sec-9">
      <title>2.3.1. Measuring completeness and correctness</title>
      <p>To compare the differences between our experimental conditions, each observation for every
participant was coded by one of the authors into three measures: a “total feature count,” “total correct
feature count,” and “total nonconforming feature count” per observation.</p>
      <p>Each observation was randomly ordered with the condition hidden, such that the coder could not tell
which observations were submitted under which conditions. The coder reviewed the content of each
submission by comparing it to the photo, counting the number of distinct pieces of information the
participant contributed about the wildlife in the photo. This process resulted in a “total feature count”
per observation. An example of this coding process, including the sample photo used, is provided in
Appendix 2.</p>
      <p>Next, the coder tallied the “total correct feature count” by comparing the text of each feature against
the photo. If the content of the observation described something that did not conform with the photo
(e.g., “furry body” for a bird, or “eight legs” for a spider with two legs obscured in the photo),
completely inferential (“looking for food” or “flies”, referring to a bird standing in a meadow),
opinionated (“nice” or “attractive”) or otherwise meaningless (“it’s difficult to say.”), it was not tallied.
All other submitted features were summed in a “total correct feature count” per observation. When in
doubt about a given feature’s uniqueness or its correctness, the coder always opted to include the
uncertain feature in the count. Last, we subtracted the total correct feature count from the total feature
count for each observation, resulting in a “total nonconforming feature count” measure for each
observation: a tally of features that were not evident by the photo alone.</p>
      <p>Does granularity influence data completeness?</p>
      <p>
        To test Hypotheses 1a and 1b, we evaluated the difference between conditions in terms of the total
correct feature count. Observations-per-condition sample sizes were low (ranging from 6 to 22)—recall
that participants had to complete at least five observations, but could quit anytime between five and 20.
We believe that some images were less likely to be seen than others due to chance distribution (images
were presented in random order). Because of these small sample sizes, we first tested the assumption
of normality using Shapiro-Wilk, finding that normality is violated for 21 cases within the 60
conditionsample pairs (tables available upon request). Since our data does not adhere to a normal distribution,
we used the Kruskal-Wallis H-test [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] to assess if the number of reported features is significantly
different between conditions. If the Kruskal-Wallis test is significant, it indicates that at least one of the
samples in a comparison is dominant over the other samples. If a given sample showed significance
according to Kruskal-Wallis, we then used pairwise comparisons with Bonferroni corrections to identify
which condition-pairs were significantly different from one another. Table 4 briefly presents the results
of these analyses.
      </p>
      <p>Observation
American Crow
.012*
.002*
.016*
.003*
.004*
.270
.005*
.108
.002*
.008*
.035*
.025*
.012*
.051
.000*
.000*
.016*
.043*</p>
      <p>Significant pairwise comparisons (Bonferroni-corrected</p>
      <p>Mann-Whitney U test p value)</p>
      <p>N/A</p>
      <p>Ungranulated–Fine granularity (.17)
Coarse granularity–Fine granularity (.036)</p>
      <p>Ungranulated–Fine granularity (.10)</p>
      <p>Ungranulated–Fine granularity (.002)
Coarse granularity–Fine granularity (.016)</p>
      <p>Ungranulated–Fine granularity (.014)</p>
      <p>Ungranulated–Fine granularity (.002)
Coarse granularity–Fine granularity (.029)
Coarse granularity–Fine granularity (.008)</p>
      <p>Ungranulated–Fine granularity (.012)</p>
      <p>N/A
Ungranulated–Fine granularity (.005)</p>
      <p>N/A
Ungranulated–Fine granularity (.001)</p>
      <p>Ungranulated–Fine granularity (.007)
Coarse granularity–Fine granularity (.036)
Coarse granularity–Fine granularity (.048)</p>
      <p>Ungranulated–Fine granularity (.033)</p>
      <p>Ungranulated–Fine granularity (.016)
Coarse granularity–Fine granularity (.032)</p>
      <p>N/A
Ungranulated–Fine granularity (.000)</p>
      <p>Ungranulated–Fine granularity (.000)
Coarse granularity–Fine granularity (.021)</p>
      <p>Ungranulated–Fine granularity (.020)
Coarse granularity–Fine granularity (.049)</p>
      <p>Ungranulated–Fine granularity (.036)</p>
      <p>As illustrated by the results in the table, we found differences between the conditions in 16/20 of the
images. In all 16 of these cases, fine granularity was consistently dominant. Moreover, in eight of these
16 cases, fine granularity observations dominated both coarse and ungranulated observations. This
evidence leads us to accept hypothesis (1a). However, our results do not show that coarse granularity
leads to a more complete dataset than ungranulated, so we reject hypothesis (1b).</p>
      <p>Does granularity influence data correctness?</p>
      <p>To test hypotheses (2a) and (2b), we examined the difference between conditions in terms of the
total nonconforming feature count for each observation. Again, we began by testing the assumption of
normality with Shapiro-Wilk (Appendix 3), and again, very few (only four) image-condition pairs
followed normal distribution. Therefore, we used the Kruskal-Wallis H-test on this data to test for
statistically significant differences between conditions in each sample. When significant differences
were found, we used Mann-Whitney pairwise comparisons with Bonferroni corrections to identify
which condition-pairs were significantly different from one another.</p>
      <p>Significant pairwise comparisons (Bonferroni-corrected</p>
      <p>Mann-Whitney U test significance value)</p>
      <p>N/A
Ungranulated–coarse granularity (.006)</p>
      <p>N/A
Fine granularity–coarse granularity (.001)
Ungranulated–coarse granularity (.034)
Fine granularity–coarse granularity (.001)
Ungranulated–coarse granularity (.009)
Fine granularity–coarse granularity (.010)
Ungranulated–coarse granularity (.006)
Fine granularity–coarse granularity (.001)
Ungranulated–coarse granularity (.001)
Ungranulated–coarse granularity (.004)
Fine granularity–coarse granularity (.002)
Ungranulated–coarse granularity (.011)</p>
      <p>N/A
N/A</p>
      <p>N/A
Ungranulated–coarse granularity (.004)
Ungranulated–coarse granularity (.028)
Ungranulated–coarse granularity (.007)
Ungranulated–coarse granularity (.000)
Fine granularity–coarse granularity (.041)
Fine granularity–coarse granularity (.001)
Ungranulated–coarse granularity (.001)</p>
      <p>N/A</p>
      <p>Ungranulated–coarse granularity (.023)</p>
      <p>As can be seen in Table 5, 14/20 images had significant differences between conditions in terms of
the number of nonconforming features contributed by participants. However, contrary to our
hypotheses, it was not the ungranulated condition that produced the most nonconforming data. Instead,
observations in the coarse granularity condition were consistently more nonconforming. This leads us
to reject hypotheses (2a) and (2b).</p>
      <p>Does granularity influence contribution quantity?</p>
      <p>Hypotheses (3a) and (3b) concern the number of contributions made by each user. We expect that
participants in ungranulated condition will provide fewer contributions than those in the granulated
conditions. To test this hypothesis, we used a one-way ANOVA to compare the means of the number
of completed observations across participants in the three conditions. Surprisingly, we found no
statistically significant difference between the groups (F(2, 84) = .773, p = .465). Therefore, we reject
hypotheses (3a) and (3b).
2.4.</p>
    </sec>
    <sec id="sec-10">
      <title>Discussion</title>
      <p>In all but four images, the number of correct features described by participants in the fine-grained
condition was substantially greater than those described by participants in the ungranulated condition.
Two of the four exceptions were photos of a bat and a whale. In both photos, the wildlife was relatively
obscure, and many participants struggled to identify what it was. The obscurity may have limited what
participants could comment on, even when asked many detailed questions about the wildlife (as in the
fine-grained condition.) In the other two exceptions, the cause of the discrepancy is less obvious. One
sample was a close-up photo of a Mourning cloak butterfly, the other of a White-throated sparrow.
Perhaps the level of detail and variety of patterns and colours visible on both subjects facilitated more
descriptive contributions.</p>
      <p>
        Otherwise, however, the evidence overwhelmingly supports the acceptance of hypothesis 1a.
Finegrained data collection interfaces drastically changed the degree of completeness of description
provided by participants. Moreover, the fine-grained data collected here was often specific to the exact
animal in the photo participants were observing (e.g., “small antlers” vs. “antlers,” in the case of the
Moose). Additionally, while it was beyond the scope of the present study to explore the usefulness of
this data, anecdotally the data includes many examples where finer-grained data may have helped
recover otherwise-bad contributions. For two particularly illustrative examples, when participants
identified a Moose as a “forest donkey” and a Harbor seal as a “sea cow”), the other features they
provided may still have been useful: according to these participants, the “forest donkey” had strong,
long legs, while the “sea cow” lived in snowy water, and in the North. In other words, when asked to
contribute granulated data participants seemed to describe the instance, not just the class, of what they
were observing [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ].
      </p>
      <p>Surprisingly, fine-grained data was substantially more detailed than coarse-grained data, while the
difference between coarse-grained data and the ungranulated condition was not statistically significant
(i.e., hypothesis 1b was rejected). It could be that there is a threshold for the benefits of fine granularity
in data collection interface designs: only after participants are encouraged to contribute a certain degree
of detail does the effect on completeness start to show. On the other hand, granulation in data collection
interface designs could have a linear relationship with completeness. A future experiment could test the
nature of this relationship with more fine-grained manipulations of granularity in a data collection
interface.</p>
      <p>
        Our second set of hypotheses was not supported by the evidence in our experiment: more
nonconforming data was introduced by participants in the coarse-grained condition, not in the
ungranulated condition as we had expected. To restate, we used Wand and Wang’s [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] notion of data
correctness to code participants’ descriptions, which notes that data that is incorrect is that which “does
not conform to [the real-world things] used to create the data” (emphasis added). If a participant stated
that the moose has “large antlers,” but the moose in the photo had short, stubby antlers, we coded this
as nonconforming. Yet, obviously, the participant was basing their contribution on their understanding
of moose in general, not the moose in the photo. The interpretation of this data as “incorrect” therefore
may be overly strict. Recoding the data to differentiate between “errors” (e.g., describing the moose as
a “hippopotamus,” as one participant did) and these generalized inferences may lead to new insights
about the relationship between granularity and data correctness.
      </p>
      <p>Still, the finding that participants using a coarse granularity interface performed the worst with this
definition of correctness is worth discussion. Perhaps these participants felt tasked to provide
information, but without the specific, detailed scope of a fine-grained interface, they weren’t sure what
else to add—thus, they provided information that wasn’t supported by the photo.</p>
      <p>It is also worth noting that there were three images in which the fine-grained condition participants
did not provide any nonconforming information—the nonconforming feature total was 0 for all
participants. The sample size of these three groups was small: 6, 7, and 10 participants. This means that
fine-grained data collection interfaces both increase data completeness (hypothesis 1a) while
maximizing data correctness (at least, defined as data that conforms to the real-world it was created
from).</p>
      <p>
        We found no statistically significant difference between groups in the number of observations
participants completed. As per hypotheses 3a and 3b, we expected that more fine-grained data collection
would be more demanding of participants, leading to fewer observations per participant in the
coarsegrained and fine-grained conditions. Participants in the fine-grained condition generally produced more
features than their counterparts in the other conditions, but perhaps fine-grained data collection made
this task easier for them to do, balancing the amount of detail against lower cognitive effort per feature.
Put another way, it may be easier to answer simple questions about the specific colours and features of
wildlife than to answer more ambiguous questions about what a photo contains. Note, however, that we
also found that participants in the fine-grained condition spent longer at the task than those in the
coarsegrained condition. Managing crowd motivation is a crucial issue [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]: asking too much of contributors
could have caused disengagement. Yet, in our experiment, participants using a fine-grained data
collection interface submitted more information, more correctly, and spent longer doing so than their
peers.
      </p>
    </sec>
    <sec id="sec-11">
      <title>3. Implications 3.1. Contributions</title>
      <p>The key contribution of this paper is considering data granularity as a key consideration for designers
of data crowdsourcing projects. In many contexts in which the observed phenomena are fleeting, data
simply cannot be re-collected. We conducted an empirical study on the effects of granularity on data
collection via crowdsourcing, finding evidence suggesting that fine-grained data facilitates the
collection of more complete and potentially more correct information, while having no effect on the
number of contributions participants were willing to make. To underscore this point: finer-grained data
may provide a more complete and correct representation of contributor observations, with no effect on
level of participation. We encourage designers of crowdsourcing platforms to strive to collect more
fine-grained data when possible, as data captured in this form may be more valuable than more
coarsegrained forms.
3.2.</p>
    </sec>
    <sec id="sec-12">
      <title>Limitations</title>
      <p>While we asked participants to self-report several important background characteristics, such as
biology/ecology education and degree of expertise in wildlife, we did not constrain these measures to
any particular geography, and we did not ask about participants’ geographical backgrounds. Participants
may have been participating from all over the world. It is possible that participants in certain conditions
were more or less familiar with wildlife in Canada’s East coast than others; future experiments should
account for this geographical factor.</p>
      <p>There are several possible limitations about our encoding of features from participant contributions.
First, only one coder encoded the collected data. To make our results more robust, we are in the process
of engaging a second coder to establish interrater reliability as part of our coding process. As previously
discussed, we may have had too strict a definition of correctness: some descriptors offered by
participants were true of the wildlife sample in general, but not evident in the photo. Another question
is whether a class-based descriptor should count as one feature, or if instead it implies many features.
Participants in the fine granularity condition still reported classes but added many features. Also,
reporting the species was not strictly the task: describing the individual animal itself was. What if
researchers were searching this labelled dataset for instances of sickly-looking animals, to surveil for
potential zoonoses? A fine-grained dataset would likely be more useful in this case (several participants
described wildlife in this experiment as “healthy looking” or similar). Moreover, this effect further
illustrates the benefits of fine-grained data when classes are mistakenly reported. In many cases where
a species is mistakenly reported but other details are provided, the other details are still useful
information.</p>
      <p>Finally, the experimental design provides only a very crude operationalization of granularity. In
practice, there could be many levels or degrees of granularity and we do not have insight into how
varying levels of granularity might have unanticipated effects on other outcomes, such as contributor
motivation or engagement.
3.3.</p>
    </sec>
    <sec id="sec-13">
      <title>Future directions</title>
      <p>
        The development of data science has been characterized in terms of three movements: business
intelligence and analytics 1.0, 2.0, and 3.0 [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. Data science 3.0 includes increased use of mobile sensor
data, more individualized and contextual analysis, and more human-centered and mobile data reporting
(e.g., visualization; [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ], see Table 1, p. 1169). To this end, is there a fourth wave of business
intelligence and analytics? The 4.0 movement might involve recognizing the important role data
contributors play in a data-driven world. To take advantage of this movement, data consumers and
analysts should account for data producers in the design of their information systems. This 4.0 wave
might therefore be characterized by design-centric data models calibrated to the ontology of the world
a given data project aims to represent. This means tuning for appropriate granulations—as a corollary,
other dimensions may be open to tuning as well.
      </p>
      <p>
        The guidelines in [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] include a stipulation for mechanisms that automatically reconcile the
instancebased data collected in the project with the coarse-grained features of a Target Organizational Model
for the project sponsor’s needs. Machine learning techniques such as supervised classifiers [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] may be
useful here. Such a technique might be used as an automatic reconciliation system that treats every new
contribution of sets of attributes as raw data and, simultaneously, as training data for an instance. A
recent study, for example, demonstrates the potential of machine learning classification by classifying
fine-grained crowdsourced data into more useful coarse-grained data with reasonable accuracy [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ].
Further explorations of how to use similar artificial intelligence tools to enhance the utility of
crowdsourced data is a potent area for future research.
      </p>
    </sec>
    <sec id="sec-14">
      <title>4. Conclusion</title>
      <p>The Internet, big data technologies, and other trends are rapidly unlocking new possibilities for
massive, directed collaboration: data crowdsourcing. These methods can allow data consumers to
collect data at unprecedented scales. However, if the data these activities generate is poorly captured, it
limits their potential value. We have provided a better understanding of the effect of finer-grained data
capture on data collection in crowdsourcing projects. Projects that enable their contributors to provide
finer-grained data may be better suited to leverage data at big data scales.</p>
    </sec>
    <sec id="sec-15">
      <title>5. Acknowledgements</title>
      <p>This research was partially supported by grants from The Natural Sciences and Engineering
Research Council of Canada (NSERC) and The Social Sciences and Humanities Research Council of
Canada (SSHRC).</p>
    </sec>
    <sec id="sec-16">
      <title>6. References</title>
    </sec>
    <sec id="sec-17">
      <title>7. Appendix 1 7.1. General Task Instructions</title>
      <p>For the remainder of your participation in this study, you will be asked to report what you observe
about a series of photos of wildlife. We are interested in identifying and labelling the flora and fauna
you see in each photo. Please be as descriptive as you can be—fill in everything you can. Identify
anything in the photo you think may be useful to researchers studying this wildlife.</p>
      <p>To qualify as having completed this task, you must complete at least five such observations. That is,
you must look at five different photos and report what you see for each. At any point before you
complete the fifth observation, you may choose to exit the study by closing the window. If you exit the
study in this way you will not be counted as having participated in the study and your data will not be
used.</p>
      <p>You may continue beyond five observations to complete as many as you’d like; more observations
are helpful for our research.</p>
    </sec>
    <sec id="sec-18">
      <title>Task Comprehension Check</title>
      <p>Before you continue, we must check that you understand the task. Please demonstrate your
comprehension by responding to the following questions:
1. In this study, you are reporting what you observe about:
• Space
• Wildlife
• People
• Architecture
2. In this study, you must complete a minimum of how many observations to participate?
• 2
• 10
• 1
• 5</p>
      <p>Note for reviewers: only participants who answer b. Wildlife for question 1 and d. 5 for question 2
will continue to participate in the study.
7.3.</p>
    </sec>
    <sec id="sec-19">
      <title>Participant Background Questionnaire</title>
      <p>Please respond to the following questions:
1. I am an expert in wildlife.</p>
      <p>• Strongly disagree
• Somewhat disagree
• Neither agree nor disagree
• Somewhat agree
• Strongly agree
2. At what level of education have you studied wildlife, ecology, or biology?
• I have never studied wildlife, ecology, or biology
• I have some high school education in wildlife, ecology, or biology
• I have some college or university education in wildlife, ecology, or biology
• I have a college or university degree in wildlife, ecology, or biology</p>
    </sec>
    <sec id="sec-20">
      <title>8. Appendix 2 8.1. Data Coding Instructions</title>
      <p>We have collected observations—descriptions of photos of wildlife—from participants using an
experimental interface. The interface instantiates conceptual model granularity into three levels:
ungranulated, coarse-grained, and fine-grained data. In our experiment, participants were randomly
assigned to one of these levels of granularity. They completed between five and 20 observations.</p>
      <p>To compare the differences between our experimental conditions, we are coding each observation
for every participant. Each observation is randomly ordered with the condition hidden, such that coders
cannot tell which observations were submitted under which conditions.</p>
      <p>The result of this coding process will be three measures of each observation: a “total feature count,”
“total correct feature count,” and “total nonconforming feature count” per observation.</p>
      <p>“Total feature count” is the total number of distinct features described in the text of each observation.
“Wing” is one feature; “Brown wing” is two: the animal has a “wing,” and the wing is “brown.”</p>
      <p>To code the “total correct feature count”, compare the text of each observation with the
corresponding photo. Total correct feature count is the total number of features that are concretely,
visibly present in the photo. “Four legs” would count as two correct features (the observed animal has
legs, and it has four of them) if and only if all four legs are visible in the photo. The contents of the
observation should not be counted when they describe:
• something that does not conform with the photo (e.g., “furry body” for a bird, or “eight legs”
for a spider with two legs obscured in the photo),
• is an assumption of the observer (“looking for food” or “flies”, referring to a bird standing
in a meadow),
• was opinionated (“nice” or “attractive”), or
• was otherwise meaningless (“it’s difficult to say.”)</p>
      <p>When in doubt about a given feature’s uniqueness or correctness, always opt to include the uncertain
feature in the count.</p>
      <p>The total nonconforming feature count is calculated by subtracting the total correct feature count
from the total feature count.</p>
      <p>To summarize, the coding algorithm is:
1. Review the contents of each part of the observation.
2. Count the number of distinct features described in the part. This is the “total feature count”—
write it in the coding sheet.</p>
      <p>o When uncertain about a given feature’s uniqueness, always opt to include the</p>
      <p>uncertain feature in the count.
3. Count the number of features that directly correspond to what is observable in the photo. This
is the “total correct feature count”—write it in the coding sheet.</p>
      <p>o Do not count non-conforming observations, assumptions, opinions, or otherwise
meaningless information.
o When uncertain about a given feature’s correctness, always opt to include the</p>
      <p>uncertain feature in the count.
4. Subtract the “total correct feature count” from the “total feature count.” This is the “total
nonconforming feature count”—write it in the coding sheet.
5. Repeat steps 1–4 for each part of the observation.
6. Repeat steps 1–5 for each observation.</p>
      <p>For demonstration purposes, an example of this process is on the next page. Notes on incorrect
features are provided for explanation only. You do not need to record your own notes when encoding
the data.</p>
    </sec>
    <sec id="sec-21">
      <title>9. Appendix 3</title>
      <p>1 (The antlers in the
photo are quite
stubby)
1
2</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>R.</given-names>
            <surname>Murphy</surname>
          </string-name>
          and
          <string-name>
            <given-names>J.</given-names>
            <surname>Parsons</surname>
          </string-name>
          ,
          <article-title>Capturing the Forest or the Trees: Designing for Granularity in Data Crowdsourcing</article-title>
          ,
          <source>in: Proceedings of the 53rd Hawaii International Conference on System Sciences</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>395</fpage>
          -
          <lpage>404</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>M.</given-names>
            <surname>Stonebraker</surname>
          </string-name>
          , What Does 'Big Data' Mean, Blog@CACM,
          <year>2012</year>
          . URL: https://cacm.acm.org/blogs/blog-cacm/155468-what
          <article-title>-does-big-data-mean/fulltext</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          <article-title>[3] The Four V's of Big Data, IBM Big Data &amp; Analytics Hub</article-title>
          , n.d. URL: https://www.ibmbigdatahub.com/infographic/four
          <article-title>-vs-big-data</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>A.</given-names>
            <surname>Castellanos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Lukyanenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Storey</surname>
          </string-name>
          , Modeling Observational Crowdsourcing, in: ER Forum,
          <source>Demo and Poster</source>
          <year>2020</year>
          ,
          <year>2020</year>
          . URL: http://ceur-ws.
          <source>org/</source>
          Vol-
          <volume>2716</volume>
          /paper11.pdf
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>R.</given-names>
            <surname>Lukyanenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Parsons</surname>
          </string-name>
          , Beyond Micro-Tasks:
          <article-title>Research Opportunities in Observational Crowdsourcing</article-title>
          ,
          <source>Journal of Database Management 29.1</source>
          (
          <year>2018</year>
          ):
          <fpage>1</fpage>
          -
          <lpage>22</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>R.</given-names>
            <surname>Lukyanenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wiersma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Huber</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Parsons</surname>
          </string-name>
          , G. Wachinger,
          <string-name>
            <given-names>R.</given-names>
            <surname>Meldt</surname>
          </string-name>
          ,
          <article-title>Representing Crowd Knowledge: Guidelines for Conceptual Modeling of User-generated Content</article-title>
          ,
          <source>Journal of the Association for Information Systems 18.4</source>
          (
          <year>2017</year>
          ):
          <fpage>297</fpage>
          -
          <lpage>339</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>R.</given-names>
            <surname>Lukyanenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Parsons</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wiersma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Maddah</surname>
          </string-name>
          , Expecting the Unexpected:
          <article-title>Effects of Data Collection Design Choices on the Quality of Crowdsourced User-Generated Content</article-title>
          ,
          <source>Management Information Systems Quarterly 43.2</source>
          (
          <year>2019</year>
          ):
          <fpage>634</fpage>
          -
          <lpage>647</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>J.</given-names>
            <surname>Parsons</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wand</surname>
          </string-name>
          ,
          <article-title>Emancipating Instances from the Tyranny of Classes in Information Modeling</article-title>
          ,
          <source>ACM Transactions on Database Systems 25.2</source>
          (
          <year>2000</year>
          ):
          <fpage>228</fpage>
          -
          <lpage>268</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wand</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. Y.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <article-title>Anchoring Data Quality Dimensions in Ontological Foundations</article-title>
          ,
          <source>Communications of the ACM</source>
          ,
          <volume>39</volume>
          .11 (
          <year>1996</year>
          ):
          <fpage>86</fpage>
          -
          <lpage>95</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>W. J.</given-names>
            <surname>Conover</surname>
          </string-name>
          , Practical Nonparametric Statistics, 3rd. ed., New York, NY, John Wiley &amp; Sons,
          <year>1998</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>J.</given-names>
            <surname>Rogstadius</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Kostakos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Kittur</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Smus</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Laredo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Vukovic</surname>
          </string-name>
          ,
          <article-title>An Assessment of Intrinsic and Extrinsic Motivation on Task Performance in Crowdsourcing Markets</article-title>
          ,
          <source>in: Proceedings of the 5th International Association for the Advancement of Artificial Intelligence Conference on Weblogs and Social Media</source>
          ,
          <year>2011</year>
          , pp.
          <fpage>321</fpage>
          -
          <lpage>328</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>H.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. H. L.</given-names>
            <surname>Chiang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Storey</surname>
          </string-name>
          , Business Intelligence and
          <article-title>Analytics: From Big Data to Big Impact</article-title>
          ,
          <source>MIS Quarterly 36.4</source>
          (
          <issue>2012</issue>
          ), pp.
          <fpage>1165</fpage>
          -
          <lpage>1188</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>F.</given-names>
            <surname>Provost</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Fawcett</surname>
          </string-name>
          ,
          <article-title>Introduction to Predictive Modeling: From Correlation to Supervised Segmentation</article-title>
          , USA,
          <string-name>
            <given-names>O</given-names>
            <surname>'Reilly Media</surname>
          </string-name>
          ,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>