=Paper= {{Paper |id=Vol-1178/CLEF2012wn-ImageCLEF-GarciaSecodeHerreraEt2012 |storemode=property |title=The medGIFT Group in ImageCLEFmed 2012 |pdfUrl=https://ceur-ws.org/Vol-1178/CLEF2012wn-ImageCLEF-GarciaSecodeHerreraEt2012.pdf |volume=Vol-1178 }} ==The medGIFT Group in ImageCLEFmed 2012 == https://ceur-ws.org/Vol-1178/CLEF2012wn-ImageCLEF-GarciaSecodeHerreraEt2012.pdf
    The medGIFT Group in ImageCLEFmed 2012

    Alba G. Seco de Herrera, Dimitrios Markonis, Ivan Eggel, Henning Müller

           University of Applied Sciences Western Switzerland (HES–SO)
                                 Sierre, Switzerland
                               alba.garcia@hevs.ch



       Abstract. This article presents the participation of the medGIFT group
       in ImageCLEFmed 2012. Since 2004, the group has participated in the
       medical image retrieval tasks of ImageCLEF each year. There are three
       types of tasks for ImageCLEFmed 2012: modality classification, image–
       based retrieval and case–based retrieval. The medGIFT group partici-
       pated in all three tasks. MedGIFT is developing a system named Par-
       aDISE (Parallel Distributed Image Search Engine), which is the successor
       of GIFT (GNU Image Finding Tool). The alpha version of ParaDISE was
       used to run most of the experiments in the competition.
       Results show that our approach using Bag–of–Visual–Words (BoVW),
       Bag–of–Colors (BoC) and Lucene for the image captions is the best run
       for mixed modality classification. The same approach is also the best for
       the image retrieval task in terms of bpref. In the case–based retrieval task,
       the Lucene baseline is the best run in terms of mean average precision
       (MAP). We were the only group presenting mixed and visual runs in
       these tasks.


1     Introduction

ImageCLEF is the cross–language image retrieval track1 of the Cross Language
Evaluation Forum (CLEF). ImageCLEFmed is part of ImageCLEF focusing on
medical images [1–5]. The medGIFT2 research group has participated in Im-
ageCLEFmed since 2004. MedGIFT is currently developing ParaDISE (Parallel
Distributed Image Search Engine), which is the successor of GIFT3 (GNU Im-
age Finding Tool). The alpha version of ParaDISE was used to run most of the
experiments.
    The Bag–of–Visual–Words (BOVW) and Bag–of–Colors (BOC) approaches
were used for visual retrieval and the textual baseline is based on the open source
Lucene4 system. In ImageCLEFmed 2011, the BoVW approach was applied
using the Scale Invariant Feature Transform (SIFT) [6] but color information
was not used. In 2012, there are two main novelties in our system: first, we
introduce the BoC [7] descriptor that represents local image colors. We combine
1
  http://www.imageclef.org/
2
  http://medgift.hevs.ch/
3
  http://www.gnu.org/software/gift/
4
  http://lucene.apache.org/
BoC with BoVW and textual descriptors in order to yield better results. Second,
we use training set expansion strategies since for some of the image categories
only very few annotated examples were available. This significantly improves
classification performance.
    The widely used BoVW [8] method is applied as follows: a training set of
images is chosen and a number of local descriptors are extracted from each image
of this set. The descriptors are then clustered using a clustering method (such
as k–means or DENCLUE). The centroids of the clusters are treated as visual
words that represent the specific local patterns. Hence, we have a visual–word
vocabulary describing all types of local image patterns. Local features are then
also extracted from each image in the database and mapped onto the visual–word
vocabulary. An image is then represented as a histograms of the visual–word
occurrences, the BoVW used as feature vector in the classification task. When
an image is queried, a similarity measure (such as histogram intersection) is
used to compare the query image histogram and the database image histogram,
providing a similarity score.
    The BoC representation is analogous to the BoVW representation and to the
Bag–of–Words representation of text documents. As the two methods are quite
complementary and as the representations as histograms are very similar, the
two approaches can easily be fused for medical image classification.
    In Section 2, we describe the datasets and the techniques used. We evaluate
the runs submitted to the ImageCLEFmed 2012 benchmark in Section 3. Finally,
conclusions are presented in Section 4.

2     Datasets and Techniques
This section describes the basic techniques used in ImageCLEFmed 2012 by the
medGIFT group. For the the majority of the runs, the alpha version of ParaDISE
(Parallel Distributed Image Search Engine) developed by the medGIFT group is
used. This image search engine is built on top of Hadoop [9], a MapReduce [10]
paradigm of parallel computing and the Cassandra 5 DBMS (Database Man-
agement System). This allows for computational and storage scalability. The
ParaDISE component–based architecture is designed to simplify the plugin of
different image features and representations. Apart from the 3 baseline visual
runs that were also given publicly to the participants, we submitted 10 runs (2
textual, 4 visual and 2 mixed runs) to the image–based retrieval task, 8 runs (2
textual, 4 visual and 2 mixed runs) to the case–based retrieval task and 8 runs
(4 visual and 4 mixed runs) to the modality classification task.

2.1    Image Collection
We used the database provided by ImageCLEFmed 2012. The database contains
over 300,000 images of 75’000 articles of the biomedical open access literature.
This is a subset from the PubMed Central6 database containing over one million
5
    http://cassandra.apache.org/
6
    http://www.ncbi.nlm.nih.gov/pmc/
images. This set of articles contains all articles in PubMed that are open access
but the exact copyright for redistribution varies among the journals. A more
detailed description of the ImageCLEFmed 2012 setup is given in [4].

2.2    Textual Techniques
For text retrieval the standard settings of the Apache Lucene text retrieval sys-
tem are used. The documents containing the journal texts are cleaned off their
XML elements and only the remaining text is used.
    Two indexations were done for ImageCLEF 2012: (1) the full text of all
articles was indexed on an article basis, and (2) the captions were indexed on a
figure basis. In the past it was shown that for case–based retrieval an indexation
of the full text had best results whereas for image–based retrieval the caption
text delivered much better results.

2.3    Visual Techniques
The baseline for the visual description of the image is the Bag–of–Visual–Words
(BoVW) approach. In this approach, local SIFT descriptors are extracted from
each image. Then, the descriptors of the image are quantized, assigning each
one to its nearest neighbour from a fixed set of local descriptors, called “visual
vocabulary”. The image is then represented by a histogram of the frequency
of the “visual words“. Similarity between two images can be quantified using a
distance metric to measure the distance between the two histograms.
    In our runs, the SIFT implementation in the fiji7 image processing pack-
age was used for the extraction of the local descriptors as in our participation
in ImageCLEFmed 2011. In order to create the visual vocabulary, our imple-
mentation of the density–based clustering algorithm DENCLUE [11] was used.
The reasons for this choice are the features and the nature of the data set that
needs to be clustered. The data set to be clustered is large (1,000 training im-
ages produce approximately 2’500’000 descriptors) and high dimensional (SIFT
descriptors are 128–dimensional). The DENCLUE algorithm is highly efficient
for clustering large–scale data sets, can detect arbitrarily shaped clusters and
handles outliers and noise well. Moreover, opposed to other density–based clus-
tering algorithms it performs well for high–dimensional data. However, when
using a density–based clustering algorithm care needs to be taken for data sets
containing clusters of different densities. To deal with this, the parameter ξ that
controls the significance of the candidate cluster in respect to its density was set
to zero. The same visual vocabulary used in last year’s participation [12] was
also used for the creation of the bags of visual words.
    Based on the BoVW approach, the Bag–of–Colors (BoC) [7] is an image
description technique introduced in [13]. Similarly to the BoVW, the technique
uses a color vocabulary C previously learned on a sub set of the collection to
represent the image. A color vocabulary C = {c1 , . . . , ckc }, with ci = (Li , ai , bi ) ∈
7
    http://fiji.sc/wiki/index.php/Fiji
CIELab is constructed by finding the most frequently occurring colors in each
image of a sub set of the collection. The modality classification training set was
used for the creation of the color vocabulary. The CIE (International Commission
on Illumination) 1976 L*a*b (CIELab) space was used, being a perceptually
uniform color space. CIELab is a space defined by L for luminance and a, b for
the color–opponent dimensions for chrominance [14, 15]. The BoC of an image I
is defined as a vector hBoC = {c̄1 , . . . , c̄k } such that, for each pixel pk ∈ I ∀k ∈
{1, . . . , np }, with np being the number of pixels of the image I:

                                  ∑  ∑
                                  np np
                          c̄i =             gj (pk ) ∀i ∈ {1, . . . , kc }
                                  k=1 j=1

where                     {
                              1 if dε (p, cj ) ≤ dε (p, cl ) ∀l ∈ {1, . . . , kc }
               gj (p) =                                                              (1)
                              0 otherwise
From experiments on the modality classification task of ImageCLEFmed 2012,
kc = 100 and kvw = 238 were chosen as the sizes of the color and visual vo-
cabulary, respectively. For the comparison of the images the histogram intersec-
tion [16] is used as distance measure.


2.4     Fusion Techniques

In image retrieval and classification tasks, often different features, systems and
results can be combined to deliver improved results. Moreover, multiple query
images can describe in more detail the visual characteristics to be retrieved.
In these cases a fusion strategy needs to be used. Two main fusion strategies
exist, early and late fusion. In the early fusion, the vectors of the features or the
systems are merged into a single vector. For the early fusion of multiple positive
and/or negative query images, Rocchio’s algorithm can be used:

                                      1 ∑           1                   ∑
                 q m = αq o + β            dj − γ                              dj    (2)
                                     |Dr |        |Dnr |
                                            dj ∈Dr                   dj ∈Dnr


where α, β and γ are weights, q m is the modified query, q o is the original query,
Dr is the set of relevant images and Dnr is the set of non–relevant images. In our
scenario there is a lack of non–relevant images, so only the second term of the
right part of the equation is used. This algorithm can be applied only for vectors
of the same feature spaces so it is not applicable for the fusion of different visual
features or retrieval systems, in general.
    In the late fusion, the retrieval results of the features or the systems are
fused. Two main categories of late fusion techniques exist, the score–based and
rank–based methods. In 2009, the ImageCLEF@ICPR fusion task was organized
to compare late fusion techniques using the best ImageCLEFmed visual and tex-
tual results [17]. Studies such as [18] show that combSUM (3) and combMNZ(4),
score–based methods, proposed by [19] in 1994 are robust fusion strategies. With
the data from the ImageCLEF@ICPR fusion task, combMNZ performed slightly
better than combSUM but the difference was small and not statistically signifi-
cant.
                                         ∑
                                         Nk
                          ScombSUM (i) =    Sk (i)                        (3)
                                            k=1

                        ScombMNZ (i) = F (i) ∗ ScombSUM (i)                    (4)
where F (i) is the frequency of image i being returned by one input system with
a non–zero score, and S(i) is the score assigned to image i.
   In general, rank–based fusion (RRF) worked better than score–based fusion.
The reciprocal rank fusion [20] is a simple fusion method based on ranks(5).
                                              ∑        1
                       RRF score(d ∈ D) =                                      (5)
                                                    k + r(d)
                                              r∈R

where, D is the set of documents retrieved, R is the set of rankings of the
documents and k = 60.


3     Results
This section details the techniques that were used to produce the runs for Im-
ageCLEFmed 2012 and then evaluates the runs.

3.1   Image and Case Retrieval Techniques
Two strategies were compared for the fusion of the multiple queries, early fusion
and late fusion (see Section 2.4). For the fusion of visual features and the fusion
of visual and textual systems, late fusion was applied. Score–based and rank–
based fusion techniques were compared. To summarize, fusion was used in three
cases:
 – fusing multiple visual features to produce visual runs;
 – fusing textual and visual runs to produce mixed runs;
 – fusing vectors (early fusion) of query images which belong to the same topic
   or their results (late fusion).
In the image retrieval task, as the fulltext search retrieved articles instead of
images, an article–to–image mapping was used. If an image was contained in
multiple articles, only the article with the highest score was taken into account
giving its score to the image. If multiple images were contained in the same
article, all the images received the common article’s score.
    The same strategy was applied in the case–based task, because the image
search retrieved images instead of articles. The article received the score of the
best scored image that it contained. If multiple articles contained the same image
they all received the common image’s score.
    Table 1 contains a summary of the techniques used, while Tables 2 and 3
show the details of the submitted runs.
           Table 1. Overview of the techniques used for the retrieval.

  Name                                Technique
  BoVW Content–based image search using Bag–of–Visual–Words representation
   BoC     Content–based image search using Bag–of–Colors representation
 Full text Lucene textual search, searching into the full text of the articles
 Captions   Lucene textual search, searching into the captions of images




              Table 2. Techniques used for the image–based runs.


Run ID Queries fusion        Techniques              Techniques fusion Run Type
  ir1   Reciprocal             BoVW                        n/a          Visual
  ir2   combMNZ                BoVW                        n/a          Visual
  ir3   combMNZ            BoVW + BoC                   combMNZ         Visual
  ir4   Reciprocal         BoVW + BoC                   Reciprocal      Visual
  ir5    Rocchio           BoVW + BoC                   combMNZ         Visual
  ir6    Rocchio           BoVW + BoC                   Reciprocal      Visual
  ir7       n/a               Full text                    n/a          Textual
  ir8       n/a               Captions                     n/a          Textual
  ir9   combMNZ BoVW + BoC + Full text + Captions       combMNZ         Mixed
 ir10   Reciprocal BoVW + BoC + Full text + Captions    Reciprocal      Mixed




               Table 3. Techniques used for the case–based runs.


Run ID Queries fusion        Techniques              Techniques fusion Run Type
  cr1   Reciprocal             BoVW                        n/a          Visual
  cr2   combMNZ                BoVW                        n/a          Visual
  cr3   combMNZ            BoVW + BoC                   combMNZ         Visual
  cr4   Reciprocal         BoVW + BoC                   Reciprocal      Visual
  cr5    Rocchio           BoVW + BoC                   combMNZ         Visual
  cr6    Rocchio           BoVW + BoC                   Reciprocal      Visual
  cr7       n/a               Full text                    n/a          Textual
  cr8       n/a               Captions                     n/a          Textual
  cr9   combMNZ BoVW + BoC + Full text + Captions       combMNZ         Mixed
 cr10   Reciprocal BoVW + BoC + Full text + Captions    Reciprocal      Mixed
3.2   Modality Classification Techniques

Driven by the good performance of [3] in the ImageCLEFmed 2011 modality
classification task, a similar approach was followed by the medGIFT group in
2012. The approach involves automatically expanding the labelled training set
to improve the performance of the classification for classes that are poorly repre-
sented (e.g. the training set containing very few images of that class). To achieve
this, training images are used as queries in the full 300,000 images of the Image-
CLEFmed 2012 data set and the l highest ranked retrieved images are added as
training images into the class of the query image.
    Two methods of expanding the training set were used. In the first expansion
technique, s images taken randomly of each class were used as queries. As in
the original training set the number of images per class varies from 5 to 50,
by choosing s = 5, l = 20 we can theoretically obtain a relatively balanced
training set (105–150 per class) of 4,100 images. In the second, all the training
images were queried, resulting in a larger non–balanced training set. E. g. for
l = 20 an expanded training set of 21,000 images can be obtained theoretically.
In practice, smaller sizes were obtained, mainly because of two reasons: retrieved
images that are already contained in the training set were discarded and images
retrieved multiple times by query images of different classes were discarded as
well.
    For a run to qualify as visual, we considered that the expanded training set
used in this run needs to be created only by visual means. This means that the
queries on the full data set used only visual features for the retrieval. Similarly,
this was repeated using mixed (visual and textual) queries for the mixed runs.
This resulted in a final number of 5 training sets (2 balanced, 2 non–balanced
and the original training set).
    A k − nn classifier using weighted voting was used to classify the test images.
For the choice of the classifier parameters the results of [7] were taken into
account and k = 11, k = 7 were used for the visual runs. However, since the
non–balanced expanded training set was significantly larger, double the value
k = 14 was also tested for this case. The inverse of the similarity score of the
k − nn images was used to weight the voting.
    Table 4 gives the details of the submitted runs.


3.3   Image Retrieval Evaluation

Table 5 displays the results of the medGIFT runs for the image retrieval task.
ir9, a mixed approach, achieved the highest MAP, GM–MAP and bpref among
all our submitted runs in the image–based retrieval task. This result highlights
the potential benefits of combining textual and visual features.
    For the visual runs, the baseline of BoVW, ir1 and ir2, did not demonstrate
good results. However, when BoVW is fused with BoC the results improve sig-
nificantly. Furthermore, when using RRF for the fusion of techniques the MAP
decreases, which contradicts our hypothesis.
                  Table 4. Modality classification runs

Run ID    Techniques       Fusion Rule   Training Set    k Run Type
 mc1        BoVW               n/a          original     11 Visual
 mc2     BoVW + BoC         combMNZ         original      7 Visual
 mc3     BoVW + BoC         combMNZ visual non-balanced 7 Visual
 mc4     BoVW + BoC         combMNZ visual non-balanced 14 Visual
 mc5     BoVW + BoC         combMNZ     visual balanced   7 Visual
 mc6 BoVW + BoC + Captions Reciprocal       original      7 Mixed
 mc7 BoVW + BoC + Captions Reciprocal mixed non-balanced 7 Mixed
 mc8 BoVW + BoC + Captions Reciprocal mixed non-balanced 14 Mixed
 mc9 BoVW + BoC + Captions Reciprocal   mixed balanced    7 Mixed




                    Table 5. Image retrieval results

        Run ID Run type MAP GM–MAP bpref        P10    P30
         best visual run 0,0101 0,0004 0,0193 0,0591 0,0439
          ir1     Visual 0,0016    0   0,0048 0,0273 0,0318
          ir2     Visual 0,0017    0   0,0058 0,0227 0,0318
          ir3     Visual 0,0049 0,0003 0,0138 0,0364 0,0364
          ir4     Visual 0,004  0,0002 0,0103 0,0227 0,0318
          ir5     Visual 0,0033 0,0003 0,0133 0,0364 0,0333
          ir6     Visual 0,003  0,0001  0,01 0,0273 0,0227
        best textual run 0.2182 0.082  0.2173 0.3409 0.2045
          ir7    Textual 0.1397 0.0436 0.1565 0.2227 0.1379
          ir8    Textual 0.1562 0.0424  0.167 0.3273 0.1864
         best mixed run 0.2377 0.0665 0.2542 0.3682 0.2712
          ir9     Mixed 0.2005 0.0917 0.1947 0.3091    0.2
         ir10     Mixed 0.1167 0.0383 0.1238 0.1864 0.1485
   Our best results in early precision (P10 and P30) on the other hand were
obtained with a text retrieval approach, ir8. This is in contradiction with past
results where the best overall results were often textual whereas early precision
was often better with combined or visual approaches.


3.4   Case Retrieval Evaluation

Table 6 shows the results of the medGIFT runs for the case based retrieval
task. In this task, cr8 achieved the highest MAP (0,169) among all submitted
runs. It demonstrates that a Lucene baseline still obtains very good results with
relatively low effort. MedGIFT was the only lab that submitted purely visual
and mixed runs for the case–based task. Although the results of our visual runs
are lower than textual runs, cr9, a mixed approach, performs better than the
average of all submitted runs in this task.


                       Table 6. Case based retrieval results

         Run ID     Run type     MAP GM-MAP bpref      P10    P30
            cr1       Visual     0,0008    0      0   0,0038 0,0013
            cr2       Visual     0,0016    0   0,0032 0,0038 0,0013
            cr3       Visual     0,0302 0,001  0,0293 0,0231 0,009
            cr4       Visual     0,0366 0,0014 0,0347 0,0269 0,0141
            cr5       Visual     0,0007    0      0      0   0,0013
            cr6       Visual     0,0008 0,0001 0,0007    0   0,0013
         second best textual run 0,1508 0,0322 0,1279 0,1538 0,1167
            cr7      Textual     0,169 0,0374 0,1499 0,1885 0,109
            cr8      Textual     0,0696 0,0028 0,0762 0,0962 0,0615
            cr9       Mixed      0,1017 0,0175 0,0857 0,1115 0,0679
           cr10       Mixed      0,0514 0,009  0,0395 0,0654 0,0564




    As all of the best results are from text retrieval runs it becomes clear that
visual techniques need to be used in different ways than was the case for obtaining
good results. Most likely a matching of image occurrences in articles in a more
complex way would be necessary for obtaining good results for the case–based
tasks using visual data.


3.5   Modality Classification Evaluation

Finally, Table 7 presents the classification accuracy of the submitted medGIFT
runs for the modality classification task. The runs c8, mc6, mc7 achieved the
three best accuracies in the mixed run category. The visual runs achieved an
average performance with the inclusion of BoC as a global descriptor to im-
proving the classification accuracy. It can be observed that the runs mc4, mc8
using the non–balanced expanded training sets and k = 14 are outperforming
                      Table 7. Modality classification results

                                   Visual                 Mixed
         Run ID       mc1 mc2 mc3 mc4 mc5 best run mc6 mc7 mc8 mc9
         Accuracy (%) 11.1 38.1 41.8 42.2 34.2 69.7 64.2 63.6 66.2 58.8



the runs mc6, mc2 that use the original training set. These runs also perform
better than the runs mc3, mc7 that use k = 7, confirming our hypothesis that
using a larger k can improve results. Moreover, in experiments not submitted
as official runs using larger values of k leads to a better performance, with the
mixed run reaching an accuracy of 68.5% using the triple value k = 21.


4   Conclusions
This article describes the methods and results of the the medGIFT group for
the ImageCLEF 2012 medical tasks. We submitted ten runs each for the ad–
hoc image retrieval and the modality classification tasks and nine runs for the
case–based retrieval task.
    In ImageCLEFmed 2012 we concentrated on fusion methods of the visual
and textual features and training set expansion strategies. We included the BoC
approach, which significantly improves classification performance.
    In the image–based retrieval task, our strategy leads to limited results ob-
tained by our submitted runs. However, we still need to do a more detailed
analysis to try to understand what is hurting performance in these runs. Our
Lucene baseline achieves the best MAP for the case–based retrieval task. Fi-
nally, in the modality classification task we submitted the three runs with the
best accuracies in the mixed run category and with only one group delivering
better results with a visual approach.
    Future work will to go beyond in the work of fusion methods, as well as
further research in new visual features. Visual information can be used in better
ways, for example to classify images by modality and if modality names appear
in a query then filtering by this modality. If the same modalities occur in a query
and an article this can also give evidence for a higher relevance.


5   Acknowledgments
The research leading to these results has received funding from the European
Union’s Seventh Framework Programme under grant agreement 257528 (KHRES-
MOI), 249008 (Chorus+) and 258191 (Promise).


References
 1. Clough, P., Müller, H., Sanderson, M.: The CLEF cross–language image retrieval
    track (ImageCLEF) 2004. In Peters, C., Clough, P., Gonzalo, J., Jones, G.J.F.,
    Kluck, M., Magnini, B., eds.: Multilingual Information Access for Text, Speech and
    Images: Result of the fifth CLEF evaluation campaign. Volume 3491 of Lecture
    Notes in Computer Science (LNCS)., Bath, UK, Springer (2005) 597–613
 2. Müller, H., Deselaers, T., Kim, E., Kalpathy-Cramer, J., Deserno, T.M., Clough,
    P., Hersh, W.: Overview of the ImageCLEFmed 2007 medical retrieval and an-
    notation tasks. In: CLEF 2007 Proceedings. Volume 5152 of Lecture Notes in
    Computer Science (LNCS)., Budapest, Hungary, Springer (2008) 473–491
 3. Kalpathy-Cramer, J., Müller, H., Bedrick, S., Eggel, I., Seco de Herrera, A.G.,
    Tsikrika, T.: The CLEF 2011 medical image retrieval and classification tasks. In:
    Working Notes of CLEF 2011 (Cross Language Evaluation Forum). (September
    2011)
 4. Müller, H., Seco de Herrera, A.G., Kalpathy-Cramer, J., Fushman, D.D., Antani,
    S., Eggel, I.: Overview of the imageclef 2012 medical image retrieval and classica-
    tion tasks. In: Working Notes of CLEF 2012 (Cross Language Evaluation Forum).
    (September 2012)
 5. Müller, H., Clough, P., Deselaers, T., Caputo, B., eds.: ImageCLEF – Experi-
    mental Evaluation in Visual Information Retrieval. Volume 32 of The Springer
    International Series On Information Retrieval. Springer, Berlin Heidelberg (2010)
 6. Lowe, D.G.: Distinctive image features from scale-invariant keypoints. Interna-
    tional Journal of Computer Vision 60(2) (2004) 91–110
 7. Seco de Herrera, A.G., Markonis, D., Müller, H.: Bag–of–colors for biomedical
    document image classification. In: MICCAI workshop MCBR-CDS. (September
    2012) Forthcoming
 8. Sivic, J., Zisserman, A.: Video google: A text retrieval approach to object match-
    ing in videos. In: Proceedings of the Ninth IEEE International Conference on
    Computer Vision - Volume 2. ICCV ’03, Washington, DC, USA, IEEE Computer
    Society (2003) 1470–1477
 9. White, T.: Hadoop: The Definitive Guide. O’Reilly Media, Inc. (2010)
10. Dean, J., Ghemawat, S.: MapReduce: simplified data processing on large clusters.
    Communications of the ACM — 50th anniversary issue 51 (January 2008) 107–113
11. Hinneburg, A., Keim, D.A.: An efficient approach to clustering in large multimedia
    databases with noise. In: Conference on Knowledge Discovery and Data Mining
    (KDD). Volume 5865., AAAI Press (1998) 58–65
12. Markonis, D., Eggel, I., Seco de Herrera, A.G., Müller, H.: The medGIFT group
    in ImageCLEFmed 2011. In: Working Notes of CLEF 2011. (2011)
13. Wengert, C., Douze, M., Jégou, H.: Bag-of-colors for improved image search. In:
    Proceedings of the 19th ACM international conference on Multimedia. MM ’11,
    New York, NY, USA, ACM (2011) 1437–1440
14. Sharma, G., Trussell, H.J.: Digital color imaging. IEEE Transactions on Image
    Processing 6(7) (1997) 901–932
15. Banu, M., Nallaperumal, K.: Analysis of color feature extraction techniques for
    pathology image retrieval system, IEEE (2010)
16. Swain, M.J., Ballard, D.H.: Color indexing. International Journal of Computer
    Vision 7(1) (1991) 11–32
17. Müller, H., Kalpathy-Cramer, J.: The ImageCLEF medical retrieval task at icpr
    2010 — information fusion to combine viusal and textual information. In: Proceed-
    ings of the International Conference on Pattern Recognition (ICPR 2010). Lecture
    Notes in Computer Science (LNCS), Istanbul, Turkey, Springer (August 2010)
18. Zhou, X., Depeursinge, A., Müller, H.: Information fusion for combining visual
    and textual image retrieval. In: 20th IEEE International Conference on Pattern
    Recognition (ICPR). (August 2010) 1590–1593
19. Fox, E.A., Shaw, J.A.: Combination of multiple searches. In: Text REtrieval
    Conference. (1993) 243–252
20. Cormack, G.V., Clarke, C.L.A., Büttcher, S.: Reciprocal rank fusion outperforms
    condorcet and individual rank learning methods. In: Proceedings of the 32nd
    international ACM SIGIR conference on Research and development in information
    retrieval, New York, NY, USA, ACM (2009) 758–759