<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>The MediaEval 2016 Context of Experience Task: Recommending Videos Suiting a Watching Situation</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Michael Riegler</string-name>
          <email>g@simula.no</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Concetto Spampinato</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Martha Larson</string-name>
          <email>m.a.larson@tudelft.nl</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Pål Halvorsen</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Carsten Griwodz</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Delft University of Technology and Radboud Univeristy</institution>
          ,
          <country country="NL">Netherlands</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Simula Research Laboratory and University of Oslo</institution>
          ,
          <country country="NO">Norway</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>University of Catania</institution>
          ,
          <country country="IT">Italy</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2016</year>
      </pub-date>
      <fpage>20</fpage>
      <lpage>21</lpage>
      <abstract>
        <p>The Context of Experience Task at MediaEval 2016 is devoted to recommending multimedia content suiting a watching situation. Speci cally, the task addresses the situation of viewers watching movies on an airplane. The goal of the task is to use trailer-content and textual metadata in order to estimate whether movies are tting to watch in ight, as judged by the crowd. The context of an airplane often falls short of an ideal movie-watching situation (noise, lack of space, interruptions, stale air, stress from turbulence) and the device can also impact user experience (small screens, glare, poor audio quality). The task explores the notion that some movies are generally better suited to these conditions than others, and that a component of this suitability is independent of viewers' personal preferences.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. INTRODUCTION</title>
      <p>The Context of Experience Task at the Multimedia
Evaluation (MediaEval) 2016 Benchmark tackles the challenge
of predicting the multimedia content that users nd most
tting to watch in speci c viewing situations. When
researchers in the area of recommender systems or multimedia
information retrieval consider the situations in which viewers
consume multimedia content, such as movies, they generally
assume comfortable watching conditions. This assumption
is understandable, since people do frequently enjoy movies
in the quiet, privacy and comfort of their own living rooms,
together with friends and loved ones, relaxing in arm chairs
and on the couch. However, movie watching is certainly not
limited to such situations. In fact, people might choose to
watch movies exactly because they are in an uncomfortable,
stressful situation and would bene t from distraction.</p>
      <p>Our ultimate goal is to build recommender systems that
support people in nding content that helps them through
tough times, i.e., moments at which they are under
psychological stress or in physical discomfort. We envisage such
contexts to include dentist o ces and hospitals. However,
here, we focus our e ort on on a context that does not
involve either physical pain or extreme psychological distress:
we chose the context of air travel. Speci cally, the
Context of Experience Task requires participants to use features
derived from video content and from movie metadata in
order to predict movies that are appropriate to watch on an
airplane.</p>
      <p>
        The next sections of the paper cover related work, and
provide more details on in- ight-distractors in uencing viewer
experience. We close with a brief description of the data
set and the task. The description is brief since this
information has been provided in detail elsewhere. Speci cally,
the rst description was published in a short paper in the
proceedings of MediaEval 2015 [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], which served to launch
the task. Additional information was published in [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ].
Finally, in order to stimulate cross-benchmark collaboration,
the task was also o ered as part of the Joint Contest on
Multimedia Challenges Beyond Visual Analysis at ICPR 2016,
and a paper published that contains a short description and
some insights on results [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ].
2.
      </p>
    </sec>
    <sec id="sec-2">
      <title>RELATED WORK</title>
      <p>
        Although our ultimate aim is to provide viewers with
multimedia content for a particular context, we di er from
context-aware movie recommendations as addressed by [
        <xref ref-type="bibr" rid="ref8 ref9">8,
9</xref>
        ]. Context of Experience assumes that the experience of
viewing a movie interacts with the context in which a movie
is viewed. Instead, we admit that a movie is actually able
to change the viewer's perception of the context. We
emphasize that addressing the challenge of recommending for
users' Contexts of Experience means not `just' matching
movies with users' personal taste, but rather also helping
users accomplish goals that they want to achieve by
consuming movies. These goals may include distracting
themselves from discomfort and making time pass more quickly.
We also note that the focus of recommender system research
on personalization often leads to neglect of cases in which
context might have a strong impact on preference relatively
independently of the personal tastes of speci c viewers, an
idea echoed in [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. Particularly strong in uence of context
can be expected in the stressful situations that are the focus
of our interest.
      </p>
      <p>
        Context of Experience is obviously closely linked to the
area of Quality of Experience of multimedia content. In [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ],
Physical context, Social Cultural Context and Task are all
identi ed as context-related factors that contribute to the
user's perception of quality of experience.
      </p>
      <p>
        Within the MediaEval benchmark1, the Context of
Experience Task follows upon other tasks that have been devoted
to predicting the impact of content on viewers or listeners.
These include an A ect Task on predicting viewer
experienced boredom [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ], the Emotion in Music task [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], a current
task on the a ective impact of movies [
        <xref ref-type="bibr" rid="ref11 ref2">11, 2</xref>
        ], and a current
task on Predicting Media Interestingness [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ].
1http://www.multimediaeval.org/
(a) The ideal situation while watching a movie on a plane.
(b) A ight attendant serving the neighboring passenger.
      </p>
      <p>(c) The movie is stopped for an announcement.
(d) Glare on the screen makes it almost impossible to see what
is going on.</p>
    </sec>
    <sec id="sec-3">
      <title>MOVIES ON A PLANE</title>
      <p>On a plane, we assume that movie viewers share the
common goal, which we consider to be a viewing intent, of
relaxing, passing time and keeping themselves occupied while
being con ned in the small and often crowded space of an
airplane cabin.</p>
      <p>Figure 1 provides an impression of a screen commonly
used on an airplane and some situations that can occur
during a ight that can in uence the watching experience of the
viewers. Sub gure 1(a) shows the optimal situation without
a distraction and a acceptable video quality. The other
subgures illustrate distracters that impact the movie viewing
experience. These examples illustrate how a person's
experience of a movie during the ight can be heavily in uenced
by the context.
4.</p>
    </sec>
    <sec id="sec-4">
      <title>TASK AND DATA</title>
      <p>The objective of the task is to classify each movie as either
+goodonairplane or -goodonairplane. Task participants are
asked to form their own hypothesis about what they think
is important for people viewing movies on an airplane, and
then to design an approach using appropriate features and
a classi er or decision function.</p>
      <p>The task data set consists of a list of movies, including
links to descriptions and video trailers, pre-extracted
features and metadata. Movies were collected between
February and April 2015 from movie lists of a major international
airline, i.e., KLM Royal Dutch Airlines. The set contains
an equal number of non-airline movies, sampled with
similar distributional properties (e.g., year). We do not
provide video les for the trailers because of copyright
restrictions. The pre-extracted visual features are Histogram of
Oriented Gradients (HOG) gray, Color Moments, local
binary patterns (LBP) and Gray Level Run Length Matrix.
The audio descriptors are Mel-Frequency Cepstral Coe
cients (MFCCs). Task participants are also allowed to
collect their own data such as full length movies, and more
metadata, e.g., user comments. The development set
contains 95 and the test set contains 223 movies. The data set is
balanced 50/50 between +goodonairplane/-goodonairplane.
The ground truth consists of user judgments gathered on
CrowdFlower. In total, 548 di erent workers participated
and at least ve judgments per movie were collected.</p>
      <p>
        For the evaluation, we use the metrics precision, recall
and weighted F1 score. We chose these metrics instead of
error rate because the task is related to recommendation.
For the purposes of recommendation, a ranked list is often
needed. Also, recall is an interesting and important part of
the evaluation. A baseline was created using a simple tree
based classi er (precision 0:629; recall of 0:573; F1 score
0:6). As mentioned above, more information is available in
the other papers that have been published discussing the
data set and the task [
        <xref ref-type="bibr" rid="ref4 ref6 ref7">7, 6, 4</xref>
        ]. We hope that the Context
of Experience Task can help to raise awareness of the topic
and also provide an interesting and meaningful use case to
inspire more work in this area.
      </p>
    </sec>
    <sec id="sec-5">
      <title>ACKNOWLEDGMENT</title>
      <p>This work is partly funded by the FRINATEK project
"EONS" (#231687) and the BIA project PCIe (#235530)
funded by the Norwegian Research Council and by the EC
FP7 project CrowdRec (#610594).</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>A.</given-names>
            <surname>Aljanaki</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.-H.</given-names>
            <surname>Yang</surname>
          </string-name>
          , and
          <string-name>
            <given-names>M.</given-names>
            <surname>Soleymani</surname>
          </string-name>
          . Emotion in Music Task:
          <article-title>Lessons learned</article-title>
          .
          <source>In this proceedings.</source>
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>E.</given-names>
            <surname>Dellandrea</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Baveye</surname>
          </string-name>
          ,
          <string-name>
            <surname>M.</surname>
          </string-name>
          <article-title>Sjoberg, and</article-title>
          <string-name>
            <given-names>C.</given-names>
            <surname>Chamaret</surname>
          </string-name>
          .
          <article-title>The MediaEval 2016 Emotional Impact of Movies Task</article-title>
          .
          <source>In this proceedings.</source>
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <surname>C.-H. Demarty</surname>
            , M. Sjoberg,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Ionescu</surname>
            , T.-T. Do,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>N. Q. K.</given-names>
          </string-name>
          <string-name>
            <surname>Duong</surname>
            , and
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Lefebvre. MediaEval 2016 Predicting Media</surname>
          </string-name>
          <article-title>Interestingness Task</article-title>
          .
          <source>In this proceedings.</source>
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>H. J.</given-names>
            <surname>Escalante</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Ponce-Lopez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Wan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. A.</given-names>
            <surname>Riegler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Clapes</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Escalera</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I.</given-names>
            <surname>Guyon</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Baro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Halvorsen</surname>
          </string-name>
          , H. Muller, and
          <string-name>
            <given-names>M.</given-names>
            <surname>Larson</surname>
          </string-name>
          .
          <article-title>ChaLearn joint contest on multimedia challenges beyond visual analysis: An overview</article-title>
          .
          <source>In Proceedings of the International Conference on Pattern Recognition</source>
          ,
          <year>2016</year>
          , to appear.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>R.</given-names>
            <surname>Pagano</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Cremonesi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Larson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Hidasi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Tikk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Karatzoglou</surname>
          </string-name>
          , and
          <string-name>
            <given-names>M.</given-names>
            <surname>Quadrana</surname>
          </string-name>
          .
          <article-title>The contextual turn: From context-aware to context-driven recommender systems</article-title>
          .
          <source>In Proceedings of the 10th ACM Conference on Recommender Systems, RecSys '16</source>
          , pages
          <fpage>249</fpage>
          {
          <fpage>252</fpage>
          ,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>M.</given-names>
            <surname>Riegler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Larson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Spampinato</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Halvorsen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Lux</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Markussen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Pogorelov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Griwodz</surname>
          </string-name>
          , and
          <string-name>
            <given-names>H.</given-names>
            <surname>Stensland</surname>
          </string-name>
          .
          <article-title>Right in ight?: A dataset for exploring the automatic prediction of movies suitable for a watching situation</article-title>
          .
          <source>In Proc. of Multimedia Systems, MMSys '16</source>
          , pages
          <issue>45:1</issue>
          {
          <issue>45</issue>
          :
          <fpage>6</fpage>
          ,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>M.</given-names>
            <surname>Riegler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Larson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Spampinato</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Markussen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Halvorsen</surname>
          </string-name>
          , and
          <string-name>
            <given-names>C.</given-names>
            <surname>Griwodz</surname>
          </string-name>
          .
          <article-title>Introduction to a task on Context of Experience: Recommending videos suiting a watching situation</article-title>
          .
          <source>In Proceedings of the MediaEval 2015 Workshop</source>
          , Wurzen, Germany,
          <source>September 14-15</source>
          ,
          <year>2015</year>
          . ceur-ws.
          <source>org/</source>
          Vol-
          <volume>1436</volume>
          /Paper5.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>A.</given-names>
            <surname>Said</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Berkovsky</surname>
          </string-name>
          , and E. W. De Luca.
          <article-title>Putting things in context: Challenge on context-aware movie recommendation</article-title>
          .
          <source>In Proceedings of the Workshop on Context-Aware Movie Recommendation</source>
          ,
          <source>CAMRa '10</source>
          , pages
          <issue>2{6</issue>
          ,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>A.</given-names>
            <surname>Said</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Berkovsky</surname>
          </string-name>
          , and E. W. De Luca.
          <article-title>Group recommendation in context</article-title>
          .
          <source>In Proceedings of the 2nd Challenge on Context-Aware Movie Recommendation</source>
          ,
          <source>CAMRa '11</source>
          , pages
          <issue>2{4</issue>
          ,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>R.</given-names>
            <surname>Schatz</surname>
          </string-name>
          , T. Ho feld, L. Janowski, and
          <string-name>
            <given-names>S.</given-names>
            <surname>Egger</surname>
          </string-name>
          . From Packets to People:
          <article-title>Quality of Experience as a New Measurement Challenge</article-title>
          , pages
          <volume>219</volume>
          {
          <fpage>263</fpage>
          . Springer Berlin Heidelberg, Berlin, Heidelberg,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>M.</given-names>
            <surname>Sjo</surname>
          </string-name>
          berg, Y. Baveye,
          <string-name>
            <given-names>H.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V. L.</given-names>
            <surname>Quang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Ionescu</surname>
          </string-name>
          , E. Dellandrea,
          <string-name>
            <given-names>M.</given-names>
            <surname>Schedl</surname>
          </string-name>
          ,
          <string-name>
            <surname>C.-H. Demarty</surname>
            , and
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Chen</surname>
          </string-name>
          .
          <article-title>The MediaEval 2015 A ective Impact of Movies Task</article-title>
          .
          <source>In Proceedings of the MediaEval 2015 Workshop Wurzen</source>
          , Germany,
          <source>September 14-15</source>
          ,
          <year>2015</year>
          . http://ceur-ws.
          <source>org/</source>
          Vol-
          <volume>1436</volume>
          /Paper1.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>M.</given-names>
            <surname>Soleymani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Larson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Pun</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <given-names>A.</given-names>
            <surname>Hanjalic</surname>
          </string-name>
          .
          <article-title>Corpus development for a ective video indexing</article-title>
          .
          <source>IEEE Transactions on Multimedia</source>
          ,
          <volume>16</volume>
          (
          <issue>4</issue>
          ):
          <volume>1075</volume>
          {
          <fpage>1089</fpage>
          ,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>