<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Report on the imageCLEF Experiment: How to visually retrieve images from the St. Andrews collection using GIFT</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Henning Muller</string-name>
          <email>henning.mueller@sim.hcuge.ch</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Antoine Geissbuhler</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Patrick Ruch</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>24 Rue Micheli-du-Crest</institution>
          ,
          <addr-line>CH-1211 Geneva 14</addr-line>
          ,
          <country country="CH">Switzerland</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>IN-Ecublens</institution>
          ,
          <addr-line>CH-1015 Lausanne</addr-line>
          ,
          <country country="CH">Switzerland</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Swiss Federal Institute of Technology</institution>
          ,
          <addr-line>LITH</addr-line>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>University and University Hospitals of Geneva, Service of Medical Informatics</institution>
        </aff>
      </contrib-group>
      <abstract>
        <p>The imageCLEF task of the Cross Language Evaluation forum has as its main goal the retrieval of images from multi{lingual test collections, or retrieval of images where the query is in a di erent language than the collection itself. The 2003 imageCLEF task saw no group using the visual information of the images that is inherently language independent. In 2004, this changed and a few groups among them the university hospitals of Geneva are submitting visual runs for the queries. The query topics are de nitely de ned in a way that makes visual retrieval extremely hard as pure visual similarity plays a marginal role whereas semantics and background knowledge are extremely important, that can only be obtained from textual captions. This article describes the submission of an entirely visual result set to the task. This article will also de ne possible improvements for visual retrieval systems with the current data. Most important is Section 4 that explains possible ways to make this query task more appealing to visual retrieval research groups, explaining problems of content{based retrieval and what such a task could do to help overcome the present problems. A benchmarking event is needed for visual information retrieval to lower current barriers in retrieval performance. ImageCLEF can help to de ne such an event and identify areas where visual retrieval might be better than textual and vice{versa. The combination of visual and textual features together is another important eld where research is needed.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>
        Visual retrieval of images has been an extremely active research area for more then ten years
now [
        <xref ref-type="bibr" rid="ref16 ref5">5, 16</xref>
        ]. Still, there has not been neither a benchmarking event nor the use of standard
datasets to compare the performance of several systems or techniques. Despite e orts such as
the Benchathlon1 [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] and several articles on evaluation [
        <xref ref-type="bibr" rid="ref11 ref12 ref17 ref8">8, 11, 12, 17</xref>
        ], no common framework has
been created, yet. This is di erent in textual information retrieval where several initiatives such
as TREC2 [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] (Text REtrieval conference) and CLEF3 [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ] (Cross Language Evaluation Forum)
exist. In 2003, CLEF added a cross language image retrieval task [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] using a collection of historic
photographs. The task in 2004 uses the same collection but adds an interactive and a medical
task [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. Figure 1 shows a few examples from the St Andrews collection.
      </p>
      <p>Images are annotated in English and query topics are formulated in another language containing
a textual description of the query and an example image. English retrieval performance is taken
(a)
(b)
(c)
(d)
as a baseline. The topics for which results can be submitted look as follows (a French example for
image 1(a)):
&lt;title&gt;
Portraits photographiques de pasteurs d'eglise par Thomas Rodger
&lt;/title&gt;
&lt;narr&gt;
Les images pertinentes sont des portraits photographiques de pasteurs ou
de leaders d'eglise pris par Thomas Ridger. Les images de nimporte quelle
epoque sont pertinentes, mais ne doivent montrer qu'une personne dans un
studio, c'est-a-dire posant pour la photo. Des photos de groupes ne sont
pas pertinentes.</p>
      <p>&lt;/narr&gt;
From this topic description we only took the image to start queries with our system, the textual
information was discarded. No manual relevance feedback or automatic query expansion was
used. This means that important information on the query task has not been obtained. With
the visual information only, we do not know that we are searching for church ministers and we
do not know who actually took the picture. Only a very good domain expert might be able to
get this information from the image alone. Actually, all this information is only ndable if the
annotation is of a very high quality and is known to be complete. It has to be assured that all
images with church ministers have these words in the text, otherwise we can not be sure whether
the person is a church minister or might have a similar function. The producer (photographer)
of the images also needs to be marked, otherwise a relevance judge would not be able to mark a
result as relevant, although two images might be extremely similar in style. What about images
where we do not have any name of the photographer but that look very similar to images from
\Thomas Ridger"? What about collections with a mediocre text quality such as those that we
often nd in the real world, for example the Internet?</p>
      <p>Some retrieval tasks led to subjectively good results with a visual retrieval system whereas
others did not manage to show any relevant images in the top 20 results. Figure 2 shows one
example result of a visual retrieval system. The rst image is the query image and we can see that
the same image was found as well as a few other images with the queen that apparently show the
same scene.</p>
      <p>Although this might look like a reasonable retrieval results, we can de nitely tell that the system
had no idea that we were looking for the queen or a military parade. The images were basically
retrieved because they have very similar properties with respect to the grey levels contained, and
especially with respect to the frame around the image. These images were most likely taken with
the same camera and digitised with the same scanner. These properties can be found with a visual
retrieval system.</p>
    </sec>
    <sec id="sec-2">
      <title>Basic technologies used for the task</title>
      <p>
        The technology used for the content{based image retrieval is mainly taken from the Viper 4 project
of the University of Geneva. Much information is available on the system [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ]. Outcome of the
Viper project is the GNU Image Finding Tool, GIFT 5. We used a version that slightly modi es
the feature space and is called medGIFT 6 as it was mainly developed for the medical domain.
These software tools are open source and can consequently also be used by other participants of
imageCLEF. Demonstration versions for participants were made available as well as not everybody
can be expected to install an entire Linux tool for such a benchmarking event, only. The feature
sets that are used by medGIFT are:
      </p>
      <p>Local colour features at di erent scales by partitioning the images successively four times
into four subregions and taking the mode colour of each region as a feature;
global colour features in the form of a colour histogram;
4http://viper.unige.ch
5http://www.gnu.org/software/gift/
6http://www.sim.hcuge.ch/medgift/
local texture features by partitioning the image and applying Gabor lters in various scales
and directions. Gabor responses are quantised into 10 strengths;
global texture features represented as a simple histogram of the responses of the local Gabor
lters in various directions and scales and with various strengths.</p>
      <p>A particularity of GIFT is that it uses many techniques from text retrieval. Visual features are
quantised/binarysed, and open a feature space that is very similar to the distribution of words in
texts (similar to a Zipf distribution). A simple tf/idf weighting is used and the query weights are
normalised by the results of the query itself. The histogram features are calculated based on a
simple histogram intersection. This allows us to apply a variety of techniques that are common
in text retrieval to the retrieval of images. Experiments show that especially relevance feedback
queries on images are much better using this feature space whereas one-shot queries might be done
more performant with other techniques.
3</p>
    </sec>
    <sec id="sec-3">
      <title>Runs submitted for evaluation</title>
      <p>Unfortunately, there was not enough time this year to submit a mixed visual and textual run for
imageCLEF but we are working on this for next year.
3.1</p>
      <sec id="sec-3-1">
        <title>Only visual retrieval with one query image</title>
        <p>For the visual queries, the medGIFT system was used. This system allows to fairly easy change a
few system parameters such as the con guration of the Gabor lters and the grey level and colour
quantisations. Input for these queries were only the query images. No feedback or automatic
query expansion was used. The following system parameters were submitted:
18 hues, 3 saturations, 3 values, 4 grey levels, 4 directions and 3 scales of the Gabor lters, the
GIFT base con guration made available to all participants of imageCLEF; (GE 4g 4d vis)
9 hues, 2 saturations, 2 values, 16 grey levels, 4 directions and 5 scales of the Gabor lters.
(GE 16g 4d vis)
Some queries delivered surprisingly good results but this was not due to a recognition of image
features with respect to the topic but rather due to the fact that images from a relevance set were
taken at a similar time and have a very similar appearance. Content{based image retrieval can
help to retrieve images that were taken with the same camera or scanned with the same scanner
if they are similar with respect to their colour properties. Mixing text and visual features for
retrieval will need a fair amount of work to optimise parameters and really receive good results.</p>
        <p>The evaluation results show the very low performance of all visual only runs that were
submitted. Mean average precision (MAP) is 0:0919 for the GIFT base system and 0:0625 for the
modi ed version. It is actually surprising that the system with only four grey levels performed
better than a system having a larger number. Most of the images are in grey and brown tones
so we expected to obtain better results when giving more exibility to this aspect. It will have
to be show whether other techniques might obtain better results such as a normalisation of the
images or even a change of the brown tones into grey tones to make images better comparable.
Still, these results will be far away from the best systems that reach a MAP of 0.5865 such as the
Daedalus system suing text retrieval only. Several systems include some visual information into
the retrieval and some of these systems are indeed ranked high. All systems that relied on visual
features, only, receive fairly bad results, in general the worst results in the competition.
3.2</p>
      </sec>
      <sec id="sec-3-2">
        <title>Techniques to improve visual retrieval results</title>
        <p>Some techniques might be of help to further increase the performance of the retrieval results.
One such techniques is a pre{processing of images to bring all images to a standard grey level
distribution and maybe removing colour completely. At least the brown levels should be changed
to grey levels so images can be retrieved based on real content and not based on general appearance.</p>
        <p>
          Another possibility is the change of the colour space of the image. Several spaces have been
analysed with respect to invariance regarding lighting conditions with good results [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ]. For the
tasks of imageCLEF it might be useful to reduce the number of colours and slightly augment the
number of grey levels for best retrieval. Some form of normalisation could also be used as some
images used the entire grey spectrum whereas others only use an extremely limited number of
grey levels. A proper evaluation will have to show what actually works best.
        </p>
        <p>Mixed visual/textual strategies can lead to a better result. If in a rst step only the textual
information is taken as a query and then the rst N images are visually fed back to the system
the results can be much better and can manage to nd images that are without text or with a
bad annotation and that would not have been found otherwise. More research is de nitely needed
on mixed textual/visual strategies for retrieval to nd out which in uence each one can have. It
might also be possible to have a small in uence of the visually most similar images in a rst query
step as well but the text will need to be the dominating factor for best results as the query topics
are semantics{based.
4</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>How to make the queries more appealing to visual retrieval research groups?</title>
      <p>Although CLEF is on cross{language retrieval and thus mainly on text, image information should
exploited in this context for the retrieval of visual data. Images are inherently language{independent
and they can provide important additional information for cross{language retrieval tasks. To
foster these developments it might even be the best to have an entirely visual task to attract the
content{based retrieval community and later come back to a combination of visual/textual
techniques. This can also help to develop partnerships between visual and textual retrieval groups to
submit common runs for such a benchmark.</p>
      <p>
        Techniques for visual information retrieval are currently not good enough to respond properly
to semantic tasks [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. Sometimes the results look indeed good but this is most often linked to
secondary parameters and not really to the semantic concepts being searched for or the low{level
features being used.
4.1
      </p>
      <sec id="sec-4-1">
        <title>More visual information for the current topics</title>
        <p>The easiest way to make the St. Andrews cross{language retrieval task more attractive to visual
retrieval groups is simply to supply more visual information as task description. Having three to
ve example images instead of one might help visual retrieval signi cantly as systems can search for
the really important information that these images have in common. A single image for retrieval
is a little bit \a shot in the dark" but several images do supply a fair amount of information.</p>
        <p>
          Besides positive examples, an important improvement would be to supply several negative
examples to have an idea of what not to look for. Negative relevance feedback has shown to
be extremely important in visual information retrieval [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ] and feedback with negative examples
substantially changes the result sets whereas positive examples only do a slight reordering of the
highest{ranked results. Finding three to ve negative examples per query task in addition to the
positive examples should not be a big problem.
4.2
        </p>
      </sec>
      <sec id="sec-4-2">
        <title>Topics based on the visual \appearance" of an image</title>
        <p>
          It has been discussed a lot what visual image retrieval cannot do but there are quite a few
things that visual image retrieval can indeed do. Although search on semantics seems currently
infeasible, similarity based on the appearance of the images can be obtained in a fairly good
quality. Visual appearance is often described as a rst impression of an image or preattentive
similarity of images [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ]. Tasks can also contain fairly easy semantics that are basically modelled
by the visual appearance. Possible topics could be:
        </p>
        <p>Sun sets { modelled by a yellow round object somewhere in the middle and mainly variations
of red.</p>
        <p>Mountain views { upper part blue and in the middle sharp changes, in grey/white tones,
bottom sometimes/often green.</p>
        <p>Beach { Lower part yellow and the upper part in blue with a clear line between the two.
City scenes { very symmetric structures with a large number of horizontal lines and right
angles.</p>
        <p>It will need to be analysed whether these queries do actually respond to what real users are
looking for in retrieval systems, but they have the potential to attract a much larger number of
visual information retrieval groups to participate and compare their techniques in such a
benchmarking event.
4.3</p>
      </sec>
      <sec id="sec-4-3">
        <title>Easy semantic topics</title>
        <p>TRECVID7 introduced in 2003 several topics for video retrieval that can also be used for visual
image retrieval, maybe with slight variations. These are fairly easy semantic topics such as nding
out whether there are people in images. Some examples for topics are:</p>
        <sec id="sec-4-3-1">
          <title>People: segment contains at least three humans.</title>
          <p>Building: segment contains a building. Buildings are walled structures with a roof.</p>
        </sec>
        <sec id="sec-4-3-2">
          <title>Road: segment contains part of a road - any size, paved or not.</title>
        </sec>
        <sec id="sec-4-3-3">
          <title>Vegetation: segment contains living vegetation in its natural environment.</title>
        </sec>
        <sec id="sec-4-3-4">
          <title>Animal: segment contains an animal other than a human .</title>
          <p>ImageCLEF could de ne topics similar in style for the image collections being available (topics
that actually do correspond to the images in the collection). Retrieval systems can then try to
nd as many of the images with respect to the topic as possible based on visual features only or
based on visual and textual features. This could also help to nd out the in uence of text and
visual information on fairly low{level semantic concepts.</p>
          <p>This can especially stimulate the creation of simple detectors for simple semantic concepts.
These detectors can later be combined for the retrieval of higher{level semantic retrieval, so they
do deliver important intermediary results.
4.4</p>
        </sec>
      </sec>
      <sec id="sec-4-4">
        <title>An easier image collection</title>
        <p>
          The St. Andrews collection is de nitely a very hard collection for purely visual analysis. The
images do not contain many clearly separated objects and the small amount of colour pictures
and variances in sharpness/quality make automatic analysis extremely hard. Other collections
such as the Corel Photo CDs are much easier for automatic analysis and query/retrieval [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ]. This
collection contains 100 images each for a large number of topics (tigers, planes, eagles, ...). Often
the collections have a distinct object in each of the sets, sometimes the sets also correspond to
regions (Paris, California, Egypt, ...). Only problem might be to get a collection without to strong
copyright constraints. As the Corel Photo CDs are not sold anymore, this might be a possibility
if Corel agrees to make the images in a lower resolution available to participants. The Corbis8
image archive also o ers a limited selection of around 15.000 images for research purposes that
are annotated in a hierarchical code. Such a collection might be an easier topic for visual and
combined visual/textual retrieval.
        </p>
        <p>
          7http://www-nlpir.nist.gov/projects/trecvid/
8http://www.corbis.com/
A di erent idea is the evaluation of interactive systems based on real users performing queries.
Normally, image retrieval is not extremely good in a rst query step but with feedback, very good
results can be obtained [
          <xref ref-type="bibr" rid="ref10 ref13">10, 13</xref>
          ]. Similar to the interactive task using text introduced in 2004
we can imagine a task with only a visual query description with an example image. Users can
subsequently perform queries until they are satis ed with the results. Evaluation could be done
directly by the users, for example by counting how many relevant images they found with which
system, and how many re nement steps were necessary to nd a satisfactory result. It has to be
stated that the user satisfaction can vary considerable with respect to his knowledge of the content
of the database. When not knowing anything about the total number of relevant images, users
tend to be satis ed fairly easily.
5
        </p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Conclusions</title>
      <p>This article explained a simple submission to the imageCLEF task using the St. Andrews historical
image collection. The two submitted runs were based on visual features of the images only, without
using the text supplied for the queries. No other techniques were used such as manual relevance
feedback or automatic query expansion. The results show the problems of purely visual image
retrieval: no semantics are currently included in the visual low{level features and as a consequence
the performance is low.</p>
      <p>Still, visual information retrieval based on low{level non{semantic features can be an important
part in the general information retrieval picture. Visual information retrieval can be used to nd
images with a similar visual appearance or with simple semantic concepts if learning data for
these concepts are available. Thus, it is important for evaluation events such as imageCLEF to
create topics that are more suitable to visual retrieval groups and that correspond to desires of real
users as well. Visual and textual retrieval need to be brought together with overlapping retrieval
tasks to nd out where each one works best and where the two can be combined for optimal
results. Currently, there is no experience in this domain, hence the importance of benchmarking
events such as imageCLEF but also the creation of retrieval tasks suitable for visual retrieval.
This article gives a few ideas on how to make the imageCLEF task more appealing for visual
retrieval groups. Hopefully, these changes will be able to attract more attention in the visual
retrieval community so people start working on the same data sets and start comparing systems
and techniques. To advance retrieval systems, a critical evaluation and comparison of existing
systems is currently more needed than new techniques. ImageCLEF might be an important factor
in advancing information retrieval and especially visual information retrieval.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>P.</given-names>
            <surname>Clough</surname>
          </string-name>
          and
          <string-name>
            <given-names>M.</given-names>
            <surname>Sanderson</surname>
          </string-name>
          .
          <article-title>The clef 2003 cross language image retrieval task</article-title>
          .
          <source>In Proceedings of the Cross Language Evaluation Forum (CLEF</source>
          <year>2004</year>
          ),
          <year>2004</year>
          (submitted).
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>P.</given-names>
            <surname>Clough</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Sanderson</surname>
          </string-name>
          , and
          <string-name>
            <given-names>H.</given-names>
            <surname>Mu</surname>
          </string-name>
          <article-title>ller. A proposal for the clef cross language image retrieval track (imageclef) 2004. In The Challenge of Image and Video Retrieval (CIVR</article-title>
          <year>2004</year>
          ), Dublin, Ireland,
          <year>July 2004</year>
          . Springer LNCS.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>D. A.</given-names>
            <surname>Forsyth</surname>
          </string-name>
          .
          <article-title>Benchmarks for storage and retrieval in multimedia databases</article-title>
          .
          <source>In Storage and Retrieval for Media Databases</source>
          , volume
          <volume>4676</volume>
          <source>of SPIE Proceedings</source>
          , pages
          <volume>240</volume>
          {
          <fpage>247</fpage>
          , San Jose, California, USA, January
          <volume>21</volume>
          {
          <fpage>22</fpage>
          2002.
          <article-title>(SPIE Photonics West Conference)</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>T.</given-names>
            <surname>Gevers</surname>
          </string-name>
          and
          <string-name>
            <given-names>A. W. M.</given-names>
            <surname>Smeulders</surname>
          </string-name>
          .
          <article-title>A comparative study of several color models for color image invariants retrieval</article-title>
          .
          <source>In Proceedings of the First International Workshop ID-MMS`96</source>
          , pages
          <fpage>17</fpage>
          {
          <fpage>26</fpage>
          , Amsterdam, The Netherlands,
          <year>August 1996</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>A.</given-names>
            <surname>Goodrum</surname>
          </string-name>
          .
          <article-title>Image information retrieval: An overview of current research</article-title>
          .
          <source>Journal of Information Science Research</source>
          ,
          <volume>3</volume>
          (
          <issue>2</issue>
          ):{,
          <year>2000</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>N. J.</given-names>
            <surname>Gunther</surname>
          </string-name>
          and
          <string-name>
            <given-names>G.</given-names>
            <surname>Beretta</surname>
          </string-name>
          .
          <article-title>A benchmark for image retrieval using distributed systems over the internet: BIRDS{I. Technical report</article-title>
          , HP Labs, Palo Alto,
          <source>Technical Report HPL{</source>
          <year>2000</year>
          {
          <fpage>162</fpage>
          , San Jose,
          <year>2001</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>D.</given-names>
            <surname>Harman</surname>
          </string-name>
          .
          <article-title>Overview of the rst Text REtrieval Conference (TREC{1)</article-title>
          .
          <source>In Proceedings of the rst Text REtrieval Conference (TREC{1)</source>
          , pages
          <fpage>1</fpage>
          <lpage>{</lpage>
          20,
          <string-name>
            <surname>Washington</surname>
            <given-names>DC</given-names>
          </string-name>
          , USA,
          <year>1992</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>C.</given-names>
            <surname>Leung</surname>
          </string-name>
          and
          <string-name>
            <given-names>H.</given-names>
            <surname>Ip</surname>
          </string-name>
          .
          <article-title>Benchmarking for content{based visual information search</article-title>
          . In R. Laurini, editor,
          <source>Fourth International Conference On Visual Information Systems</source>
          (VISUAL'
          <year>2000</year>
          ),
          <source>number 1929 in Lecture Notes in Computer Science</source>
          , pages
          <volume>442</volume>
          {
          <fpage>456</fpage>
          ,
          <string-name>
            <surname>Lyon</surname>
          </string-name>
          , France,
          <year>November 2000</year>
          . Springer{Verlag.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>H.</given-names>
            <surname>Mu</surname>
          </string-name>
          <article-title>ller, S. Marchand-Maillet, and</article-title>
          <string-name>
            <given-names>T.</given-names>
            <surname>Pun</surname>
          </string-name>
          .
          <article-title>The truth about corel { evaluation in image retrieval</article-title>
          .
          <source>In Proceedings of the International Conference on the Challenge of Image and Video Retrieval (CIVR</source>
          <year>2002</year>
          ), London, England,
          <year>July 2002</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>H.</given-names>
            <surname>Mu</surname>
          </string-name>
          ller, W. Muller,
          <string-name>
            <given-names>D. M.</given-names>
            <surname>Squire</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Marchand-Maillet</surname>
          </string-name>
          , and
          <string-name>
            <given-names>T.</given-names>
            <surname>Pun</surname>
          </string-name>
          .
          <article-title>Strategies for positive and negative relevance feedback in image retrieval</article-title>
          . In A. Sanfeliu,
          <string-name>
            <given-names>J. J.</given-names>
            <surname>Villanueva</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Vanrell</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Alcezar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.-O.</given-names>
            <surname>Eklundh</surname>
          </string-name>
          , and Y. Aloimonos, editors,
          <source>Proceedings of the 15th International Conference on Pattern Recognition (ICPR</source>
          <year>2000</year>
          ), pages
          <fpage>1043</fpage>
          {
          <fpage>1046</fpage>
          ,
          <string-name>
            <surname>Barcelona</surname>
          </string-name>
          , Spain,
          <year>September 2000</year>
          . IEEE.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>H.</given-names>
            <surname>Mu</surname>
          </string-name>
          ller, W. Muller,
          <string-name>
            <given-names>D. M.</given-names>
            <surname>Squire</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Marchand-Maillet</surname>
          </string-name>
          , and
          <string-name>
            <given-names>T.</given-names>
            <surname>Pun</surname>
          </string-name>
          .
          <article-title>Performance evaluation in content{based image retrieval: Overview and proposals</article-title>
          .
          <source>Pattern Recognition Letters</source>
          ,
          <volume>22</volume>
          (
          <issue>5</issue>
          ):
          <volume>593</volume>
          {
          <fpage>601</fpage>
          ,
          <string-name>
            <surname>April</surname>
          </string-name>
          <year>2001</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <surname>A. D. Narasimhalu</surname>
            ,
            <given-names>M. S.</given-names>
          </string-name>
          <string-name>
            <surname>Kankanhalli</surname>
            , and
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Wu</surname>
          </string-name>
          .
          <article-title>Benchmarking multimedia databases</article-title>
          .
          <source>Multimedia Tools and Applications</source>
          ,
          <volume>4</volume>
          :
          <fpage>333</fpage>
          {
          <fpage>356</fpage>
          ,
          <year>1997</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Rui</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T. S.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ortega</surname>
          </string-name>
          , and
          <string-name>
            <given-names>S.</given-names>
            <surname>Mehrotra</surname>
          </string-name>
          .
          <article-title>Relevance feedback: A power tool for interactive content{based image retrieval</article-title>
          .
          <source>IEEE Transactions on Circuits and Systems for Video Technology</source>
          ,
          <volume>8</volume>
          (
          <issue>5</issue>
          ):
          <volume>644</volume>
          {
          <fpage>655</fpage>
          ,
          <year>September 1998</year>
          .
          <article-title>(Special Issue on Segmentation, Description, and Retrieval of Video Content)</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>S.</given-names>
            <surname>Santini</surname>
          </string-name>
          and
          <string-name>
            <given-names>R.</given-names>
            <surname>Jain</surname>
          </string-name>
          .
          <article-title>Gabor space and the development of preattentive similarity</article-title>
          .
          <source>In Proceedings of the 13th International Conference on Pattern Recognition (ICPR'96)</source>
          , pages
          <fpage>40</fpage>
          {
          <fpage>44</fpage>
          , Vienna, Austria,
          <year>August 1996</year>
          . IEEE.
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>J.</given-names>
            <surname>Savoy</surname>
          </string-name>
          .
          <source>Report on clef{2001 experiments. In Report on the CLEF Conference 2001 (Cross Language Evaluation Forum)</source>
          , pages
          <fpage>27</fpage>
          {
          <fpage>43</fpage>
          ,
          <string-name>
            <surname>Darmstadt</surname>
          </string-name>
          , Germany,
          <year>2002</year>
          . Springer LNCS 2406.
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <surname>A. W. M. Smeulders</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Worring</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Santini</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Gupta</surname>
            , and
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Jain</surname>
          </string-name>
          .
          <article-title>Content{based image retrieval at the end of the early years</article-title>
          .
          <source>IEEE Transactions on Pattern Analysis and Machine Intelligence</source>
          ,
          <volume>22</volume>
          No 12:
          <fpage>1349</fpage>
          {
          <fpage>1380</fpage>
          ,
          <year>2000</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>J. R.</given-names>
            <surname>Smith</surname>
          </string-name>
          .
          <article-title>Image retrieval evaluation</article-title>
          .
          <source>In IEEE Workshop on Content{based Access of Image and Video Libraries (CBAIVL'98)</source>
          , pages
          <fpage>112</fpage>
          {
          <fpage>113</fpage>
          ,
          <string-name>
            <surname>Santa</surname>
            <given-names>Barbara</given-names>
          </string-name>
          , CA, USA, June 21
          <year>1998</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <surname>D. M. Squire</surname>
            , W. Muller, H. Muller, and
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Pun</surname>
          </string-name>
          .
          <article-title>Content{based query of image databases: inspirations from text retrieval</article-title>
          .
          <source>Pattern Recognition Letters (Selected Papers from The 11th Scandinavian Conference on Image Analysis SCIA '99)</source>
          ,
          <volume>21</volume>
          (
          <fpage>13</fpage>
          -14):
          <volume>1193</volume>
          {
          <fpage>1198</fpage>
          ,
          <year>2000</year>
          .
          <string-name>
            <given-names>B.K.</given-names>
            <surname>Ersboll</surname>
          </string-name>
          , P. Johansen, Eds.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>