<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Retrieving Diverse Social Images at MediaEval 2017: Challenges, Dataset and Evaluation</article-title>
      </title-group>
      <contrib-group>
        <aff id="aff0">
          <label>0</label>
          <institution>CEA LIST</institution>
          ,
          <country country="FR">France</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Maia Zaharieva</institution>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>TU Wien</institution>
          ,
          <country country="AT">Austria</country>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>Universidade Federal de Minas Gerais</institution>
          ,
          <country country="BR">Brazil</country>
        </aff>
        <aff id="aff4">
          <label>4</label>
          <institution>University Politehnica of Bucharest</institution>
          ,
          <country country="RO">Romania</country>
        </aff>
        <aff id="aff5">
          <label>5</label>
          <institution>University of Applied Sciences Western Switzerland</institution>
          ,
          <country country="CH">Switzerland</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2017</year>
      </pub-date>
      <fpage>13</fpage>
      <lpage>15</lpage>
      <abstract>
        <p>This paper provides an overview of the Retrieving Diverse Social Images task that is organized as part of the MediaEval 2017 Benchmarking Initiative for Multimedia Evaluation. The task addresses the challenge of visual diversification of image retrieval results, where images, metadata, user tagging profiles, and content and text models are available for processing. We present the task challenges, the employed dataset and ground truth information, the required runs, and the considered evaluation metrics.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>INTRODUCTION</title>
      <p>
        An eficient image retrieval system should be able to present results
that are both relevant to the provided query and that are covering
diferent visual aspects of it. The diversification of image search
results can considerably increase the probability of a system to
address a broad range of user information needs. In general,
diversification is an actively researched problem in various domains
ranging from web search and query result diversification [
        <xref ref-type="bibr" rid="ref15 ref19 ref9">9, 15, 19</xref>
        ]
to recommender systems [
        <xref ref-type="bibr" rid="ref16 ref17">16, 17</xref>
        ] and summarization [
        <xref ref-type="bibr" rid="ref13 ref14">13, 14</xref>
        ]. With
the emerging availability of publicly available images, the
importance for diversification of image data is steadily growing. This task
is especially challenging when handling real-world queries, which
are often complex, consisting of multiple concepts.
      </p>
      <p>
        The 2017 Retrieving Diverse Social Images task is a follow-up of
the 2016 edition [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] and fosters the development of new techniques
for improving both the relevance and the visual diversification of
image search results. The task is designed to support the evaluation
and comparison of approaches emerging from a wide range of
research fields, such as information retrieval (text, vision, multimedia
communities), machine learning, relevance feedback, and natural
language processing.
      </p>
    </sec>
    <sec id="sec-2">
      <title>TASK DESCRIPTION</title>
      <p>
        The task is built around the use case of a general ad-hoc image
retrieval system, which provides the user with visually diversified
representations of query results (see for instance Google Image
Search1). Given a ranked list of up to 300 query-related images
retrieved from Flickr2 using text-based queries, participants are
required to refine the results by providing a set of images that are
relevant to the query and, at the same time, represent a visually
diversified summary of it. The queries include complex and
generalpurpose, multi-concept queries (e.g. "dancing on the street", "trees
reflected in water", "sailing boat"). The queries in the development
set result from a broad user study and are constructed around the
the data of the MediaEval 2016 Retrieving Diverse Social Images
task [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. The queries in the test set were collected using Google
Trends3 for image search (worldwide, last 5 years: 2012-2017).
      </p>
      <p>The goal of the task is to refine the image set retrieved as a result
to a given text-based query by providing a ranked list of up to 50
photos that are both relevant and visually diversified representations
of the query, according to the following definitions:
Relevance: an image is considered to be relevant for the query
if it is a common visual representation of the query topics (all at
once). Bad quality photos (e.g., severely blurred, out of focus) are
not considered relevant in this scenario;
Diversity: a set of images is considered to be diverse if it depicts
diferent visual characteristics of the query topics and subtopics
with a certain degree of complementarity, i.e. most of the perceived
visual information is diferent from one image to another.
3</p>
    </sec>
    <sec id="sec-3">
      <title>DATA DESCRIPTION</title>
      <p>The data consists of a development set (devset) with 110 queries
(32, 487 images) and a test set (testset) with 84 queries (24, 986
images). An additional dataset (credibilityset) provides credibility
estimation for ca. 685 users and metadata for more than 3.5M
images. We also provide semantic vectors for general English terms
computed on top of the English Wikipedia4 (wikiset), which could
help participants to develop advanced text models.</p>
      <p>Each query is accompanied by the following information: query
text formulation (the actual query formulation used on Flickr to
retrieve the data), a ranked list of up to 300 images in jpeg format
retrieved from Flickr using Flickr’s default "relevance" algorithm
(all images are redistributable Creative Commons licensed5), an
1https://images.google.com/
2https://www.flickr.com.
3http://trends.google.com/
4https://en.wikipedia.org/
5http://creativecommons.org/
xml file containing Flickr metadata for the retrieved images, and
ground truth for both relevance and diversity.</p>
      <p>
        To facilitate participation from various communities, we also
provide the following content-based descriptors:
- general purpose, visual-based descriptors extracted using the LIRE
library6 [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]: auto color correlogram (ACC) [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]; color and edge
directivity descriptor (CEDD) [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], fuzzy color and texture histogram
(FCTH) [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], Gabor texture, joint composite descriptor (JCD) [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ],
several MPEG7 features including color layout, edge histogram, and
scalable color [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ], pyramid of histograms of orientation gradients
(PHOG) [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], and speeded up robust features (SURF) [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ].
- convolutional neural network (CNN)-based descriptors based on the
reference model provided with the Cafe framework 7. The
descriptors are extracted from the last fully connected layer (fc7).
- text-based features include term frequency and document
frequency information and their ratio (TF-IDF). The text-based features
are computed per image, per query, and per user basis.
- user annotation credibility descriptors provide an estimation of the
quality of the users’ tag-image content relationships. The
following descriptors are provided: visualScore (measure of user image
relevance), faceProportion (the percentage of images with faces),
tagSpecificity (average specificity of a user’s tags, where tag
speciifcity is the percentage of users having annotated with that tag in a
large Flickr corpus), locationSimilarity (average similarity between a
user’s geotagged photos and a probabilistic model of a surrounding
cell), photoCount (total number of images a user shared), uniqueTags
(proportion of unique tags), uploadFrequency (average time between
two consecutive uploads), bulkProportion (the proportion of bulk
taggings in a user’s stream, i.e., of tag sets that appear identical for
at least two distinct photos), meanPhotoViews (mean value of the
number of times a user’s image has been seen by other members of
the community), meanTitleWordCounts (mean value of the number
of words found in the titles associated with users’ photos),
meanTagsPerPhoto (mean value of the number of tags users put for their
images), meanTagRank (mean rank of a user’s tags in a list in which
the tags are sorted in descending order according the number of
appearances in a large subsample of Flickr images), and
meanImageTagClarity (adaptation of the Image Tag Clarity from [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ] using
a TF-IDF language model as individual tag language model).
4
      </p>
    </sec>
    <sec id="sec-4">
      <title>GROUND TRUTH</title>
      <p>Both relevance and diversity annotations were carried out by 17
human annotators. The data were distributed among the annotators
such that each query was labeled by three diferent annotators. For
relevance, annotators were asked to label each image (one at a time)
as being relevant to the underlying query (value 1), non-relevant
(0), or with "don’t know" (−1). The final relevance ground truth
score was determined using a majority voting scheme. For
diversity, only the images that were judged as relevant in the previous
step were considered. For each query, annotators were provided
with a thumbnail list of all relevant images. After getting familiar
with their contents, they were asked to re-group the images into
clusters with similar visual appearance (up to 25 clusters in total).
6http://www.lire-project.net/
7http://cafe.berkeleyvision.org/</p>
      <p>In contrast to the single relevance score for each query, in terms
of diversity we consider all three annotations as correct (ground
truth) as they typically depict diferent possibilities to group the
images representing diferent points of view.
5</p>
    </sec>
    <sec id="sec-5">
      <title>RUN DESCRIPTION</title>
      <p>Participants were allowed to submit up to five runs. The first three
are required (dedicated) runs: run1 – automated run using visual
information only; run2 – automated run using text information
only; and run3 – automated run using both visual and text
information. For the generation of run1 to run3 only information that
can be extracted from the provided data (e.g. provided descriptors,
descriptors of their own, etc.) is allowed to be used. The last two
runs, run4 and run5, are general ones, i.e. any approach is allowed,
e.g. human-based or hybrid human-machine approaches, including
using data from external sources, such as Internet or pre-trained
models obtained from external datasets related to this task.
6</p>
    </sec>
    <sec id="sec-6">
      <title>EVALUATION</title>
      <p>
        Performance is assessed for both diversity and relevance using
cluster recall at X (CR@X ), precision at X (P @X ), and their harmonic
mean F 1@X . CR@X provides the ratio of the number of clusters
from the ground truth that are represented in the top X results and,
thus, it reflects the diversification quality of a given image result
set. We compute CR@X for each one of the available ground truth
diversity annotations and select the one which maximizes CR@X
for each query. Since the clusters in the ground truth consider
relevant images only, the relevance of the top X results is implicitly
measured by CR@X . Nevertheless, P @X provides a more precise
view on the relevance of a particular image set since it directly
measures the relevance among the top X images. We consider
various cut of points, i.e. X = {5, 10, 20, 30, 40, 50}. Additionally, we
consider two further evaluation metrics, which are well-established
in the information retrieval community, the intent-aware expected
reciprocal rank (ERR-IA@X ) [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] and the α -normalized discounted
cumulative gain (α -nDCG@X ) [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] metrics. The oficial ranking
metric is F 1@20 which gives equal importance to diversity (via CR@20)
and relevance (via P @20). This metric simulates the content of a
single page of a typical Web image search engine and reflects user
behavior, i.e., inspecting the first page of results with priority.
7
      </p>
    </sec>
    <sec id="sec-7">
      <title>CONCLUSION</title>
      <p>The 2017 Retrieving Diverse Social Images task provides
participants with a comparative and collaborative evaluation benchmark
for social image retrieval approaches focusing on visual-based
diversification . The task explores the diversification in the context of
a challenging, ad-hoc image retrieval system, which should be able
to tackle complex and general-purpose multi-concept queries. This
year, we explicitly accounted for the possibility of having
multiple diferent views on a given retrieval result, which might all be
subjectively correct. This allows for an investigation of the aspect
of subjectivity in the perception of diversification in a next step.
Details on the methods and results of the participating teams can be
found in the working note papers of the MediaEval 2017 workshop
proceedings.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>Herbert</given-names>
            <surname>Bay</surname>
          </string-name>
          , Andreas Ess, Tinne Tuytelaars, and Luc Van Gool.
          <year>2008</year>
          .
          <article-title>Speeded-Up Robust Features (SURF)</article-title>
          .
          <source>Computer Vision and Image Understanding</source>
          <volume>110</volume>
          ,
          <issue>3</issue>
          (
          <year>2008</year>
          ),
          <fpage>346</fpage>
          -
          <lpage>359</lpage>
          . https://doi.org/10.1016/j.cviu.
          <year>2007</year>
          .
          <volume>09</volume>
          .014
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>Anna</given-names>
            <surname>Bosch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Andrew</given-names>
            <surname>Zisserman</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Xavier</given-names>
            <surname>Munoz</surname>
          </string-name>
          .
          <year>2007</year>
          .
          <article-title>Representing Shape with a Spatial Pyramid Kernel</article-title>
          . In ACM International Conference on Image and
          <article-title>Video Retrieval (CIVR)</article-title>
          . ACM, New York, NY, USA,
          <fpage>401</fpage>
          -
          <lpage>408</lpage>
          . https://doi.org/10. 1145/1282280.1282340
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>Olivier</given-names>
            <surname>Chapelle</surname>
          </string-name>
          , Donald Metlzer, Ya Zhang, and
          <string-name>
            <given-names>Pierre</given-names>
            <surname>Grinspan</surname>
          </string-name>
          .
          <year>2009</year>
          .
          <article-title>Expected Reciprocal Rank for Graded Relevance</article-title>
          .
          <source>In ACM Conference on Information and Knowledge Management (CIKM)</source>
          . ACM, New York, NY, USA,
          <fpage>621</fpage>
          -
          <lpage>630</lpage>
          . https: //doi.org/10.1145/1645953.1646033
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <surname>Savvas</surname>
            <given-names>A Chatzichristofis</given-names>
          </string-name>
          , YS Boutalis, and
          <string-name>
            <given-names>Mathias</given-names>
            <surname>Lux</surname>
          </string-name>
          .
          <year>2009</year>
          .
          <article-title>Selection of the proper compact composite descriptor for improving content based image retrieval</article-title>
          .
          <source>In Signal Processing</source>
          ,
          <article-title>Pattern Recognition and Applications (SPPRA)</article-title>
          .
          <source>ACTA Press</source>
          ,
          <volume>134</volume>
          -
          <fpage>140</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <surname>Savvas</surname>
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Chatzichristofis</surname>
            and
            <given-names>Yiannis S.</given-names>
          </string-name>
          <string-name>
            <surname>Boutalis</surname>
          </string-name>
          .
          <year>2008</year>
          .
          <article-title>CEDD: Color and Edge Directivity Descriptor: A Compact Descriptor for Image Indexing and Retrieval</article-title>
          .
          <source>In International Conference on Computer Vision Systems (ICCV)</source>
          . Springer-Verlag, Berlin, Heidelberg,
          <fpage>312</fpage>
          -
          <lpage>322</lpage>
          . https://doi.org/10.1007/978-3-
          <fpage>540</fpage>
          -79547-6_
          <fpage>30</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>S. A.</given-names>
            <surname>Chatzichristofis</surname>
          </string-name>
          and
          <string-name>
            <given-names>Y. S.</given-names>
            <surname>Boutalis</surname>
          </string-name>
          .
          <year>2008</year>
          .
          <article-title>FCTH: Fuzzy Color and Texture Histogram - A Low Level Feature for Accurate Image Retrieval</article-title>
          .
          <source>In Int. Workshop on Image Analysis for Multimedia Interactive Services (WIAMIS)</source>
          .
          <source>IEEE Computer Society</source>
          , Washington, DC, USA,
          <fpage>191</fpage>
          -
          <lpage>196</lpage>
          . https://doi.org/10.1109/WIAMIS.
          <year>2008</year>
          .24
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>Charles</surname>
            <given-names>L.A.</given-names>
          </string-name>
          <string-name>
            <surname>Clarke</surname>
          </string-name>
          , Maheedhar Kolla,
          <string-name>
            <surname>Gordon</surname>
            <given-names>V.</given-names>
          </string-name>
          <string-name>
            <surname>Cormack</surname>
          </string-name>
          , Olga Vechtomova, Azin Ashkan, Stefan Büttcher, and Ian MacKinnon.
          <year>2008</year>
          .
          <article-title>Novelty and Diversity in Information Retrieval Evaluation</article-title>
          . In
          <source>International ACM SIGIR Conference on Research and Development in Information Retrieval. ACM</source>
          , New York, NY, USA,
          <fpage>659</fpage>
          -
          <lpage>666</lpage>
          . https://doi.org/10.1145/1390334.1390446
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>Bogdan</given-names>
            <surname>Ionescu</surname>
          </string-name>
          , Alexandru Lucian Ginsca, Maia Zaharieva, Bogdan Boteanu, Mihai Lupu, and
          <string-name>
            <given-names>Henning</given-names>
            <surname>Müller</surname>
          </string-name>
          .
          <year>2016</year>
          . Retrieving Diverse Social Images at MediaEval 2016:
          <article-title>Challenge, Dataset and Evaluation</article-title>
          .
          <source>In MediaEval 2016 Multimedia Benchmark Workshop</source>
          , Vol.
          <volume>1739</volume>
          . CEUR-WS.org.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>Bogdan</given-names>
            <surname>Ionescu</surname>
          </string-name>
          , Adrian Popescu,
          <string-name>
            <surname>Anca-Livia Radu</surname>
            , and
            <given-names>Henning</given-names>
          </string-name>
          <string-name>
            <surname>Müller</surname>
          </string-name>
          .
          <year>2016</year>
          .
          <article-title>Result diversification in social image retrieval: a benchmarking framework</article-title>
          .
          <source>Multimedia Tools and Applications</source>
          <volume>75</volume>
          ,
          <issue>2</issue>
          (
          <year>2016</year>
          ),
          <fpage>1301</fpage>
          -
          <lpage>1331</lpage>
          . https://doi.org/10. 1007/s11042-014-2369-4
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>Mathias</given-names>
            <surname>Lux</surname>
          </string-name>
          .
          <year>2011</year>
          .
          <article-title>Content Based Image Retrieval with LIRe</article-title>
          .
          <source>In ACM International Conference on Multimedia. ACM</source>
          , New York, NY, USA,
          <fpage>735</fpage>
          -
          <lpage>738</lpage>
          . https://doi.org/10.1145/2072298.2072432
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>B.S.</given-names>
            <surname>Manjunath</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.-R.</given-names>
            <surname>Ohm</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.V.</given-names>
            <surname>Vasudevan</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <given-names>A.</given-names>
            <surname>Yamada</surname>
          </string-name>
          .
          <year>2001</year>
          .
          <article-title>Color and texture descriptors</article-title>
          .
          <source>IEEE Transactions on Circuits and Systems for Video Technology</source>
          <volume>11</volume>
          ,
          <issue>6</issue>
          (
          <year>2001</year>
          ),
          <fpage>703</fpage>
          -
          <lpage>715</lpage>
          . https://doi.org/10.1109/76.927424
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <surname>Mandar</surname>
            <given-names>Mitra</given-names>
          </string-name>
          , Ramin Zabih, Jing Huang,
          <string-name>
            <surname>Wei-Jing Zhu</surname>
            , and
            <given-names>S. Ravi</given-names>
          </string-name>
          <string-name>
            <surname>Kumar</surname>
          </string-name>
          .
          <year>1997</year>
          .
          <article-title>Image Indexing Using Color Correlograms</article-title>
          .
          <source>In IEEE Conference on Computer Vision</source>
          and
          <article-title>Pattern Recognition (CVPR)</article-title>
          .
          <source>IEEE Computer Society</source>
          , Washington, DC, USA,
          <fpage>762</fpage>
          -
          <lpage>768</lpage>
          . https://doi.org/10.1109/CVPR.
          <year>1997</year>
          .609412
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <surname>Yanwei</surname>
            <given-names>Pang</given-names>
          </string-name>
          , Qiang Hao, Yuan Yuan, Tanji Hu, Rui Cai,
          <string-name>
            <given-names>and Lei</given-names>
            <surname>Zhang</surname>
          </string-name>
          .
          <year>2011</year>
          .
          <article-title>Summarizing Tourist Destinations by Mining User-generated Travelogues and Photos</article-title>
          .
          <source>Computer Vision and Image Understanding</source>
          <volume>115</volume>
          ,
          <issue>3</issue>
          (
          <year>2011</year>
          ),
          <fpage>352</fpage>
          -
          <lpage>363</lpage>
          . https: //doi.org/10.1016/j.cviu.
          <year>2010</year>
          .
          <volume>10</volume>
          .010
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>S.</given-names>
            <surname>Rudinac</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Hanjalic</surname>
          </string-name>
          , and
          <string-name>
            <given-names>M.</given-names>
            <surname>Larson</surname>
          </string-name>
          .
          <year>2013</year>
          .
          <article-title>Generating Visual Summaries of Geographic Areas Using Community-Contributed Images</article-title>
          .
          <source>IEEE Transactions on Multimedia 15</source>
          ,
          <issue>4</issue>
          (
          <year>2013</year>
          ),
          <fpage>921</fpage>
          -
          <lpage>932</lpage>
          . https://doi.org/10.1109/TMM.
          <year>2013</year>
          .2237896
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <surname>Rodrygo</surname>
            <given-names>L. T.</given-names>
          </string-name>
          <string-name>
            <surname>Santos</surname>
            , Craig Macdonald, and
            <given-names>Iadh</given-names>
          </string-name>
          <string-name>
            <surname>Ounis</surname>
          </string-name>
          .
          <year>2015</year>
          .
          <article-title>Search result diversification</article-title>
          .
          <source>Foundations and Trends in Information Retrieval 9</source>
          ,
          <issue>1</issue>
          (
          <year>2015</year>
          ),
          <fpage>1</fpage>
          -
          <lpage>90</lpage>
          . https://doi.org/10.1561/1500000040
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>Markus</given-names>
            <surname>Schedl</surname>
          </string-name>
          and
          <string-name>
            <given-names>David</given-names>
            <surname>Hauger</surname>
          </string-name>
          .
          <year>2015</year>
          .
          <article-title>Tailoring Music Recommendations to Users by Considering Diversity, Mainstreaminess, and Novelty</article-title>
          . In
          <source>International ACM SIGIR Conference on Research and Development in Information Retrieval. ACM</source>
          , New York, NY, USA,
          <fpage>947</fpage>
          -
          <lpage>950</lpage>
          . https://doi.org/10.1145/2766462.2767763
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <surname>Yue</surname>
            <given-names>Shi</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>Xiaoxue</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <surname>Jun</surname>
            <given-names>Wang</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Martha Larson</surname>
            , and
            <given-names>Alan</given-names>
          </string-name>
          <string-name>
            <surname>Hanjalic</surname>
          </string-name>
          .
          <year>2012</year>
          .
          <article-title>Adaptive Diversification of Recommendation Results via Latent Factor Portfolio</article-title>
          . In
          <source>International ACM SIGIR Conference on Research and Development in Information Retrieval. ACM</source>
          , New York, NY, USA,
          <fpage>175</fpage>
          -
          <lpage>184</lpage>
          . https://doi.org/10.1145/2348283. 2348310
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>Aixin</given-names>
            <surname>Sun</surname>
          </string-name>
          and
          <string-name>
            <given-names>Sourav S.</given-names>
            <surname>Bhowmick</surname>
          </string-name>
          .
          <year>2009</year>
          .
          <article-title>Image Tag Clarity: In Search of Visualrepresentative Tags for Social Images</article-title>
          .
          <source>In SIGMM Workshop on Social Media. ACM</source>
          , New York, NY, USA,
          <fpage>19</fpage>
          -
          <lpage>26</lpage>
          . https://doi.org/10.1145/1631144.1631150
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <surname>Kaiping</surname>
            <given-names>Zheng</given-names>
          </string-name>
          , Hongzhi Wang, Zhixin Qi,
          <string-name>
            <given-names>Jianzhong</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>and Hong</given-names>
            <surname>Gao</surname>
          </string-name>
          .
          <year>2016</year>
          .
          <article-title>A survey of query result diversification</article-title>
          .
          <source>Knowledge and Information Systems</source>
          <volume>51</volume>
          ,
          <issue>1</issue>
          (
          <year>2016</year>
          ),
          <fpage>1</fpage>
          -
          <lpage>36</lpage>
          . https://doi.org/10.1007/s10115-016-0990-4
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>