<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Christina Boididou</string-name>
          <email>boididou@iti.gr</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Symeon Papadopoulos</string-name>
          <email>papadop@iti.gr</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Duc-Tien Dang-Nguyen</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Giulia Boato</string-name>
          <email>boato@disi.unitn.it</email>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Michael</string-name>
          <email>michael@simula.no</email>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Riegler</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Stuart E. Middleton</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Andreas Petlund</string-name>
          <email>apetlund@ifi.uio.no</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Yiannis Kompatsiaris</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Information Technologies Institute</institution>
          ,
          <addr-line>CERTH</addr-line>
          ,
          <country country="GR">Greece</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Simula Research Laboratory</institution>
          ,
          <country country="NO">Norway</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>University of Southampton IT Innovation Centre</institution>
          ,
          <addr-line>Southampton</addr-line>
          ,
          <country country="UK">UK</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2016</year>
      </pub-date>
      <fpage>20</fpage>
      <lpage>21</lpage>
      <abstract>
        <p>This paper provides an overview of the Verifying Multimedia Use task that takes places as part of the 2016 MediaEval Benchmark. The task motivates the development of automated techniques for detecting manipulated and misleading use of web multimedia content. Splicing, tampering and reposting videos and images are examples of manipulation that are part of the task de nition. For the 2016 edition of the task, a corpus of images/videos and their associated posts is made available, together with labels indicating the appearance of misuse (fake) or not (real) in each case as well as some useful post metadata.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. INTRODUCTION</title>
      <p>
        Social media, such as Twitter and Facebook, as means of
news sharing is very popular and also very often used by
government or politicians to reach the public. The speed of news
spreading on such platforms often leads to the appearance
of large amounts of misleading multimedia content. Given
the need for automated real-time veri cation of this content,
several techniques have been presented by researchers. For
instance, previous work focused on the classi cation between
fake and real tweets spread during Hurricane Sandy [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] and
other events [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] or on automatic methods for assessing posts'
credibility [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. Several systems for checking content
credibility have been proposed, such as Truthy [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], TweetCred
[
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] and Hoaxy [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. The second edition of this task aims to
encourage the development of new veri cation approaches.
This year, the task is extended by introducing a sub-task,
focused on identifying digitally manipulated multimedia
content. To this end, we encourage participants to create
textfocused and/or image-focused approaches equally.
      </p>
    </sec>
    <sec id="sec-2">
      <title>TASK OVERVIEW</title>
      <p>Main task. The de nition of the main task is the following:
\Given a social media post, comprising a text component,
an associated piece of visual content (image/video) and a set
of metadata originating from the social media platform, the
task requires participants to return a decision (fake, real or
unknown) on whether the information presented by this post
(a) (b) (c)
Figure 1: Examples of misleading (fake) image use: (a)
reposting of real photo claiming to show two Vietnamese
siblings at Nepal 2015 earthquake; (b) reposting of
artwork as a photo of solar eclipse (March 2015); (c) spliced
sharks on a photo during Hurricane Sandy in 2012.
su ciently re ects the reality." In practice, participants
receive a list of posts that are associated with images and are
required to automatically predict, for each post, whether it is
trustworthy or deceptive (real or fake respectively). In
addition to fully automated approaches, the task also considers
human-assisted approaches provided that they are practical
(i.e., fast enough) in real-world settings. The following
definitions should be also taken into account:</p>
      <p>A post is considered fake when it shares multimedia
content that does not represent the event that it refers to.
Figure 1 presents examples of such content.</p>
      <p>A post is considered real when it shares multimedia that
legitimately represents the event it refers to.</p>
      <p>A post that shares multimedia content that does not
represent the event it refers to but reports the false
information or refers to it with a sense of humour is neither
considered fake nor real (and hence not included in
the task dataset).</p>
      <p>Sub-task. This version of the task addresses the problem
of detecting digitally manipulated (tampered) images. The
de nition of the task is the following: \Given an image, the
task requires participants to return a decision (tampered,
non-tampered or unknown) on whether the image has been
digitally modi ed or not". In practice, participants receive
a list of images and are required to predict if this image
is tampered or not, using multimedia forensic analysis. It
should also be noted that an image is considered tampered
when it is digitally altered.</p>
      <p>In both cases, the task also asks participants to optionally
return an explanation (which can be a text string, or URLs
pointing to evidence) that supports the veri cation decision.
The explanation is not used for quantitative evaluation, but
rather for gaining qualitative insights into the results.</p>
    </sec>
    <sec id="sec-3">
      <title>VERIFICATION CORPUS</title>
      <p>
        Development dataset (devset): This is provided together
with ground truth and is used by participants to develop
their approach. For the main task, it contains posts related
to the 17 events of Table 1, comprising in total 193 cases
of real and 220 cases of misused images/videos, associated
with 6,225 real and 9,596 fake posts posted by 5,895 and
9,216 unique users respectively. This data is the union of
last year's devset and testset [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. Note that several of the
events, e.g., Columbian Chemicals and Passport Hoax are
hoaxes, hence all multimedia content associated with them
is misused. For several real events (e.g., MA ight 370) no
real images (and hence no real posts) are included in the
dataset, since none came up as a result of the data collection
process that is described below. For the sub-task, the
development set contains 33 cases of non-tampered and 33 cases
of tampered images, derived from the same events, along
with their labels (tampered and non-tampered).
Test dataset (testset): This is used for evaluation. For
the main task, it comprises 104 cases of real and misused
images and 25 cases of real and misused videos, in total
associated with 1,107 and 1,121 posts, respectively. For the
sub-task, it includes 64 cases of both tampered and
nontampered images from the testset events.
      </p>
      <p>
        The data for both datasets are publicly available1.
Similar to the 2015 edition of the task, the posts were collected
around a number of known events or news stories and
contain fake and real multimedia content manually veri ed by
cross-checking online sources (articles and blogs). Having
de ned a set of keywords K for each testset event, we
collected a set of posts P (using Twitter API and speci c
keywords) and a set of unique fake and real pictures around
these events, resulting in the fake and real image sets IF , IR
respectively. We then used the image sets as seeds to
create our reference veri cation corpus PC P , which includes
only those posts that contain at least one image of the
prede ned sets IF , IR. However, in order not to restrict the
posts to the ones pointing to the exact image, we employed
a scalable visual near-duplicate search strategy [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]: we used
the IF , IR as visual queries and for each query we checked
whether each post image from the P set exists as an image
item or a near-duplicate image item of the IF or the IR set.
In addition to this process, we also used a real-time system
that collects posts using keywords and a location lter [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ].
This was performed mainly to increase the real samples for
events that occurred in known locations.
      </p>
      <p>To further extend the testset, we carried out a
crowdsourcing campaign using the microWorkers platform2. We
asked each worker to provide three cases of manipulated
multimedia content that they found on the web. Furthermore,
they had to provide a link with information and description
on each case, along with online resources containing evidence
of its misleading nature. We also asked them to provide the
original content if available. To avoid cheating, they had
to provide a manual description of the manipulation. We
also tested the task in two pilot studies to be sure that the
1https://github.com/MKLab-ITI/image-verification-corpus/
tree/master/mediaeval2016
2https://microworkers.com/
information we got would also be useful. Overall, the data
collected was very useful. We performed 75 tasks and each
worker earned 2; 75$ per task.</p>
      <p>
        For every item of the datasets, we extracted and made
available three types of features, similar to the ones we made
available for the 2015 edition of the task: (i) features
extracted from the post itself, i.e., the number of words,
hashtags, mentions, etc. in the post's text [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], (ii) features
extracted from the user account, i.e., number of friends
and followers, whether the user is veri ed, etc. [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. and
(iii) forensic features extracted from the image, i.e.,
the probability map of the aligned double JPEG
compression, the estimated quantization steps for the rst six DCT
coe cients of the non-aligned JPEG compression, and the
Photo-Response Non-Uniformity (PRNU) [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ].
4.
      </p>
    </sec>
    <sec id="sec-4">
      <title>EVALUATION</title>
      <p>Overall, the main task is interested in the accuracy with
which an automatic method can distinguish between use of
multimedia in posts in ways that faithfully re ect reality
versus ways that spread false impressions. Hence, given a
set of labelled instances (post + image + label) and a set of
predicted labels (included in the submitted runs) for these
instances, the classic IR measures (i.e., Precision P , Recall
R, and F -score) are used to quantify the classi cation
performance, where the target class is the class of fake tweets.
Since the two classes (fake/real) are represented in a
relatively balanced way in the testset, the classic IR measures
are good proxies of the classi er accuracy. Note that task
participants are allowed to classify a tweet as unknown.
Obviously, in case a system produces many unknown outputs,
it is likely that its precision will bene t, assuming that the
selection of unknown is done wisely, i.e. successfully
avoiding erroneous classi cations. However, the recall of such a
system will su er in case the tweets that are labelled as
unknown turn out to be fake (the target class). Similarly, in
the sub-task case, given the instances of (image + label), we
use the same IR measures to quantify the performance of
the approach, where the target class is tampered.</p>
    </sec>
    <sec id="sec-5">
      <title>ACKNOWLEDGEMENTS</title>
      <p>This work is supported by the REVEAL and InVID projects,
partially funded by the European Commission (FP7-610928 and
H2020-687786 respectively).</p>
    </sec>
    <sec id="sec-6">
      <title>REFERENCES</title>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>C.</given-names>
            <surname>Boididou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Andreadou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Papadopoulos</surname>
          </string-name>
          ,
          <string-name>
            <surname>D.-T.</surname>
            Dang-Nguyen,
            <given-names>G.</given-names>
          </string-name>
          <string-name>
            <surname>Boato</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Riegler</surname>
            , and
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>Kompatsiaris</surname>
          </string-name>
          .
          <article-title>Verifying multimedia use at mediaeval 2015</article-title>
          .
          <source>In Working Notes Proceedings of the MediaEval 2015 Workshop</source>
          ,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>C.</given-names>
            <surname>Boididou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Papadopoulos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Kompatsiaris</surname>
          </string-name>
          ,
          <string-name>
            <surname>S.</surname>
          </string-name>
          <article-title>Schi eres, and</article-title>
          <string-name>
            <given-names>N.</given-names>
            <surname>Newman</surname>
          </string-name>
          .
          <article-title>Challenges of computational veri cation in social multimedia</article-title>
          .
          <source>In Proceedings of the 23rd International Conference on World Wide Web</source>
          , pages
          <volume>743</volume>
          {
          <fpage>748</fpage>
          . ACM,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>C.</given-names>
            <surname>Castillo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Mendoza</surname>
          </string-name>
          , and
          <string-name>
            <given-names>B.</given-names>
            <surname>Poblete</surname>
          </string-name>
          .
          <article-title>Information credibility on twitter</article-title>
          .
          <source>In Proceedings of the 20th international conference on World wide web</source>
          , pages
          <volume>675</volume>
          {
          <fpage>684</fpage>
          . ACM,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>V.</given-names>
            <surname>Conotter</surname>
          </string-name>
          ,
          <string-name>
            <surname>D.-T.</surname>
            Dang-Nguyen,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Riegler</surname>
            , G. Boato, and
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Larson</surname>
          </string-name>
          .
          <article-title>A crowdsourced data set of edited images online</article-title>
          .
          <source>In Proceedings of the 2014 International ACM Workshop on Crowdsourcing for Multimedia</source>
          ,
          <source>CrowdMM '14</source>
          , pages
          <fpage>49</fpage>
          {
          <fpage>52</fpage>
          , New York, NY, USA,
          <year>2014</year>
          . ACM.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>A.</given-names>
            <surname>Gupta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Kumaraguru</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Castillo</surname>
          </string-name>
          , and
          <string-name>
            <given-names>P.</given-names>
            <surname>Meier</surname>
          </string-name>
          . Tweetcred:
          <article-title>Real-time credibility assessment of content on twitter</article-title>
          .
          <source>In International Conference on Social Informatics</source>
          , pages
          <volume>228</volume>
          {
          <fpage>243</fpage>
          . Springer,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>A.</given-names>
            <surname>Gupta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Lamba</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Kumaraguru</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <given-names>A.</given-names>
            <surname>Joshi</surname>
          </string-name>
          .
          <article-title>Faking Sandy: characterizing and identifying fake images on twitter during Hurricane Sandy</article-title>
          .
          <source>In Proceedings of the 22nd international conference on World Wide Web companion</source>
          , pages
          <volume>729</volume>
          {
          <fpage>736</fpage>
          ,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>S. E.</given-names>
            <surname>Middleton</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Middleton</surname>
          </string-name>
          , and
          <string-name>
            <surname>S.</surname>
          </string-name>
          <article-title>Moda eri. Real-time crisis mapping of natural disasters using social media</article-title>
          .
          <source>IEEE Intelligent Systems</source>
          ,
          <volume>29</volume>
          (
          <issue>2</issue>
          ):9{
          <fpage>17</fpage>
          ,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>J.</given-names>
            <surname>Ratkiewicz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Conover</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Meiss</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Goncalves</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Patil</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Flammini</surname>
          </string-name>
          , and
          <string-name>
            <given-names>F.</given-names>
            <surname>Menczer</surname>
          </string-name>
          .
          <article-title>Truthy: mapping the spread of astroturf in microblog streams</article-title>
          .
          <source>In Proceedings of the 20th international conference companion on World wide web</source>
          , pages
          <volume>249</volume>
          {
          <fpage>252</fpage>
          . ACM,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>C.</given-names>
            <surname>Shao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G. L.</given-names>
            <surname>Ciampaglia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Flammini</surname>
          </string-name>
          , and
          <string-name>
            <given-names>F.</given-names>
            <surname>Menczer</surname>
          </string-name>
          .
          <article-title>Hoaxy: A platform for tracking online misinformation</article-title>
          .
          <source>In Proceedings of the 25th International Conference Companion on World Wide Web</source>
          , pages
          <volume>745</volume>
          {
          <fpage>750</fpage>
          . International World Wide Web Conferences Steering Committee,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>E.</given-names>
            <surname>Spyromitros-Xiou s</surname>
          </string-name>
          , S. Papadopoulos, I. Kompatsiaris, G. Tsoumakas,
          <string-name>
            <surname>and I. Vlahavas.</surname>
          </string-name>
          <article-title>A comprehensive study over VLAD and Product Quantization in large-scale image retrieval</article-title>
          .
          <source>IEEE Transactions on Multimedia</source>
          ,
          <volume>16</volume>
          (
          <issue>6</issue>
          ):
          <volume>1713</volume>
          {
          <fpage>1728</fpage>
          ,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>