<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Overview of the CLEF 2010 medical image retrieval track</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Henning Mu¨ller</string-name>
          <email>henning.mueller@sim.hcuge.ch</email>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Jayashree Kalpathy-Cramer</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ivan Eggel</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Steven Bedrick</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Joe Reisetter</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Charles E. Kahn Jr.</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>William Hersh</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Radiology, Medical College of Wisconsin</institution>
          ,
          <addr-line>Milwaukee, WI</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Geneva University Hospitals and University of Geneva</institution>
          ,
          <country country="CH">Switzerland</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Oregon Health and Science University (OHSU)</institution>
          ,
          <addr-line>Portland, OR</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>University of Applied Sciences Western Switzerland</institution>
          ,
          <addr-line>Sierre</addr-line>
          ,
          <country country="CH">Switzerland</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>The seventh edition of the ImageCLEF medical retrieval task was organized in 2010. As in 2008 and 2009, the collection in 2010 uses images and captions from the Radiology and Radiographics journals published by RSNA (Radiological Society of North America). Three subtasks were conducted within the auspices of the medical task: modality detection, image-based retrieval and case-based retrieval. The goal of the modality detection task was to detect the acquisition modality of the images in the collection using visual, textual or mixed methods. The goal of the image-based retrieval task was to retrieve an ordered set of images from the collection that best met the information need specified as a textual statement and a set of sample images, while the goal of the case-based retrieval task was to return an ordered set of articles (rather than images) that best met the information need provided as a description of a “case”. The number of registrations to the medical task increased to 51 research groups. However, groups submitting runs have remained stable at 16, with the number of submitted runs increasing to 155. Of these, 61 were ad-hoc runs, 48 were case-based runs while the remaining 46 were modality classification runs. The best results for the ad-hoc retrieval topics were obtained using mixed methods with textual methods also performing well. Textual methods were clearly superior for the case-based topics. For the modality detection task, although textual and visual methods alone were relatively successful, combining these techniques proved most effective.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        ImageCLEF1 [
        <xref ref-type="bibr" rid="ref1 ref2 ref3">1–3</xref>
        ] started in 2003 as part of the Cross Language Evaluation
Forum (CLEF2, [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]). A medical image retrieval task was added in 2004 and has
been held every year since [
        <xref ref-type="bibr" rid="ref3 ref5">3, 5</xref>
        ]. The main goal of ImageCLEF continues to be
      </p>
    </sec>
    <sec id="sec-2">
      <title>1 http://www.imageclef.org/</title>
      <p>2 http://www.clef-campaign.org/
promoting multi–modal information retrieval by combining a variety of media
including text and images for more effective information retrieval.</p>
      <p>In 2010, the format of CLEF was changed from a workshop at the European
Conference on Digital Libraries (ECDL) to an independent conference on
multilingual and multimedia retrieval evaluation3 which includes several organized
evaluation tasks now called labs.
2</p>
      <sec id="sec-2-1">
        <title>Participation, Data Sets, Tasks, Ground Truth</title>
        <p>This section describes the details concerning the set–up and the participation in
the medical retrieval task in 2010.
2.1</p>
        <sec id="sec-2-1-1">
          <title>Participation</title>
          <p>In 2010, a new record of 112 research groups registered for the four sub–tasks
of ImageCLEF down from seven sub tasks in 2009. For the medical retrieval
task the number of registrations also reached a new maximum with 51. 16 of the
participants submitted results to the tasks, essentially the same number as in
previous years. The following groups submitted at least one run:
– AUEB (Greece);
– Bioingenium (Columbia)∗;
– Computer Aided Medical Diagnoses (Edu??),∗;
– Gigabioinforamtics (Belgium)∗;
– IRIT (France);
– ISSR (Egypt);
– ITI, NIH (USA);
– MedGIFT (Switzerland);
– OHSU (USA);
– RitsMIP (Japan)∗;
– Sierre, HES–SO (Switzerland);
– SINAI (Spain);
– UAIC (Romania)∗;
– UESTC (China)∗;
– UIUC–IBM (USA)∗;
– Xerox (France)∗.</p>
          <p>Participants marked with a star had never before participated in the medical
retrieval task, indicating that the number of first–time participants was fairly
high with eight among the 16 participants.</p>
          <p>A total of 155 valid runs were submitted, 46 of which were submitted for
modality detection, 61 for the image–based topics and 48 for the case–based
topics. The number of runs per group was limited to ten per subtask and case–
based and image–based topics were seen as separate subtasks in this view.</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3 http://www.clef2010.org/</title>
      <p>2.2</p>
      <sec id="sec-3-1">
        <title>Datasets</title>
        <p>The database used in 2009 was again made accessible by the Radiological Society
of North America (RSNA4). The database contained a total of 77,506 images,
and was the largest collection to ever have been used for ImageCLEFmed. All
images in the collection originated from the journals Radiology and Radiographics,
published by the RSNA. A similar database is also available via the Goldminer5
interface. This collection constitutes an important body of medical knowledge
from the peer–reviewed scientific literature including high quality images with
textual annotations. Images are associated with journal articles, and can also
be part of a larger figure. Figure captions were made available to participants,
as well as the sub–caption concerning a particular subfigure (if available). This
high–quality set of textual annotations enabled textual searching in addition to
content–based retrieval. Furthermore, the PubMed IDs of each figure’s
originating article were also made available, allowing participants to access the MeSH
(Medical Subject Headings) index terms assigned by the National Library of
Medicine for MEDLINE6.
2.3</p>
      </sec>
      <sec id="sec-3-2">
        <title>Modality Classification</title>
        <p>
          Previous research [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] has demonstrated the utility of classifying images by
modality in order to improve the precision of the search. The modality classification
task was conceived as the first step for the medical image retrieval task whereby
participants use the modality classifier created in this step to improve their
performance for the retrieval task. For this task, 2390 images were provided as a
training set where each image was classified as belonging to one of 8 classes (CT,
GX, MR, NM, PET, PX, US, XR). One of the authors (JKC) had manually,
but somewhat cursorily, verified the assigned modality of all images. 2620 test
images were provided for the task. Each of these images were to be assigned
a modality using visual, textual or mixed techniques. Participants were also
requested to provide a classification for all images in the collection. A majority vote
classification for all images in the collection was made available upon request to
participants of the task after the evaluation.
2.4
        </p>
      </sec>
      <sec id="sec-3-3">
        <title>Image–Based Topics</title>
        <p>
          The topics for the image-based retrieval task were created using methods similar
to previous years where realistic search topics were identified by surveying actual
user needs. The starting point for the 2010 topics was a user study [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ] conducted
at Oregon Health &amp; Science University (OHSU) in early 2009. Using qualitative
methods, this study was conducted with medical practitioners and was focused
on understanding their needs, both met and unmet, regarding medical image
4 http://www.rsna.org/
5 http://goldminer.arrs.org/
6 http://www.pubmed.gov/
retrieval. The first part of the study was dedicated to the investigation of the
demographics and characteristics of participants, a population served by medical
image retrieval systems (e.g., their background, searching habits, etc.). After a
demonstration of state–of–the–art image retrieval systems, the second part of
the study was devoted to learning about the motivation and tasks for which
the intended audience uses medical image retrieval systems (e.g., contexts in
which they seek medical images, types of useful images, numbers of desired
answers, etc.). In the third and last part, the participants were asked to use the
demonstrated systems to try to solve challenging queries, and provide responses
to questions investigating how likely they would be to use such systems, aspects
they did and did not like, and missing features they would like to see added.
In total, the 37 participants utilized the demonstrated systems to perform a
total of 95 searches using textual queries in English. We randomly selected 25
candidate queries from the 95 searches to create the topics for ImageCLEFmed
2009. Similarly, this year, we randomly selected another 25 queries from the
remaining queries. From these, using the OHSU image retrieval system which
was indexed using the 2009 ImageCLEF collection, we finally selected 16 topics
for which at least one relevant image was retrieved by the system.
        </p>
        <p>We added 2 to 4 sample images to each query from the previous collections
of ImageCLEFmed. Then, for each topic, we provided a French and a German
translation of the original textual description provided by the participants.
Finally, the resulting set of topics was categorized into three groups: 3 visual topics,
9 mixed topics, and 4 semantic topics. This categorization of topics was based
on the organizers’ prior experience with how amenable certain types of search
topics are to visual, textual or mixed search techniques. However, this is not an
exact science and was merely provided for guidance. The entire set of topics was
finally approved by a physician.
2.5</p>
      </sec>
      <sec id="sec-3-4">
        <title>Case–Based Topics</title>
        <p>Case–based topics were made available for the first time in 2009, and in 2010
the number of case–based topics was increased from 5 to 14, roughly half of all
topics. The goal was to move image retrieval potentially closer to clinical routine
by simulating the use case of a clinician who is in the process of diagnosing a
difficult case. Providing this clinician with articles from the literature that discuss
cases similar7 to the case (s)he is working on can be a valuable aid to choosing
a good diagnosis or treatment.</p>
        <p>
          The topics were created based on cases from the teaching file Casimage [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ].
This teaching file contains cases (including images) from radiological practice
that clinicians document mainly for using them in teaching. 20 cases were pre–
selected and a search with the diagnosis was performed in the ImageCLEF data
set to make sure that there were at least a few matching articles. Fourteen topics
were finally chosen. The diagnosis and all information on the chosen treatment
was then removed from the cases so as to simulate the situation of the clinician
7 “Similar” in terms of images and other clinical data on the patient.
who has to diagnose the patient. In order to make the judging more consistent,
the relevance judges were provided with the original diagnosis for each case.
2.6
        </p>
      </sec>
      <sec id="sec-3-5">
        <title>Relevance Judgements</title>
        <p>The relevance judgements were performed with the same on–line system as in
2008 and 2009 for the image–based topics as well as case–based topics. The
system had been adapted in 2009 for the case–based topics, displaying the article
title and several images appearing in the text (currently the first six, but this
can be configured). Judges were provided with a protocol for the process with
specific details on what should be regarded as relevant versus non–relevant. A
ternary judgement scheme was used again, wherein each image in each pool was
judged to be “relevant”, “partly relevant”, or “non–relevant”. Images clearly
corresponding to all criteria were judged as “relevant”, images whose relevance
could not be accurately confirmed but could still be possible were marked as
“partly relevant”, and images for which one or more criteria of the topic were
not met were marked as “non–relevant”. Judges were instructed in these criteria
and results were manually verified during the judgement process. As in previous
years, judges were recruited by sending out an e–mail to current and former
students at OHSU’s Department of Medical Informatics and Clinical Epidemiology.
Judges, primarily clinicians, were paid a small stipend for their services. Many
topics were judged by two or more judges to explore inter–rater agreements and
its effects on the robustness of the rankings of the systems.
3</p>
        <sec id="sec-3-5-1">
          <title>Results</title>
          <p>This section describes the results of ImageCLEF 2010. Runs are ordered based on
the techniques used (visual, textual, mixed) and the interaction used (automatic,
manual). Case–based topics and image–based topics are separated but compared
in the same sections. Trec eval was used for the evaluation process, and we made
use of most of its performance measures.
3.1</p>
        </sec>
      </sec>
      <sec id="sec-3-6">
        <title>Submissions</title>
        <p>The numbers of submitting teams was slightly lower in 2010 than in 2009 with 16
instead of 17. The numbers of runs increased from 124 to 155. The distribution
among the three run types of modality detection, image–based retrieval and case–
based retrieval showed that all three types reached almost the same number of
submissions.</p>
        <p>Groups subsequently had the chance to evaluate additional runs themselves
as the qrels were made available to participants two weeks ahead of the
submission deadline for the working notes.
3.2</p>
      </sec>
      <sec id="sec-3-7">
        <title>Modality Detection Results</title>
        <p>A variety of commonly used image processing techniques were explored by the
participants. Features used included local binary patterns (LBP) [9], Tamura
texture features [10], Gabor features [11], the GIFT (GNU Image Finding Tool),
the Color Layout Descriptor (CLD) and Edge Histogram Descriptor (EHD) from
MPEG–7, Color and Edge Directivity Descriptor (CEDD) and Fuzzy Color and
Texture Histogram (FCTH) using the Lucene image retrieval (LIRE) library,
Scale Invariant Feature Transform (SIFT) [12] as well as various combinations
of these. Classifiers ranged from simple k–nearest neighbors (kNN) to Ada–
Boost, multilayer perceptrons and Support Vector Machines (SVMs) as well as
a variety of techniques to combine the output from multiple classifiers including
those derived from Bayes theory such as product, sum, maximum and mean
rules</p>
        <p>The results of the modality detection tasks are given in Table 1below. As
seen in the table, the best results were obtained using mixed methods (94%)
for the modality classification task. The best run using textual methods (90%)
had a slightly better accuracy than the best run using visual methods (87%).
However, for groups that submitted runs using different methods, the best results
were obtained when they combined visual and textual methods.
3.3</p>
      </sec>
      <sec id="sec-3-8">
        <title>Image–Based Retrieval Results</title>
        <p>The best results for the ad–hoc retrieval topics were obtained using mixed
methods. Textual methods, as in previous years also performed well. However, visual
methods by themselves, were not very effective for this collection.
Visual Retrieval As in previous years, only 8 of the 61 submitted runs used
purely visual techniques. As discussed previously, this collection, with extremely
well annotated textual captions and images that are primarily from radiology,
does not lend itself to purely visual techniques. However, as seen from the results
of the mixed runs, the use of the visual information contained in the image can
improve the search performance over that of a purely textual system.</p>
        <p>An analysis of the results shows that most techniques are in a very similar
range and only a single run had a significantly better result in terms of MAP.
The baseline system GIFT (GNU Image Finding Tool) is in the upper half of
the performance.</p>
        <p>Textual Retrieval Participants explored a variety of information retrieval
techniques from the use of stop word removal and stemming to utilizing Lucene or
Lemur, commonly used toolkits to techniques using Latent Semantic Indexing,
database searches using full–text Boolean queries, query expansion with
external sources such as MeSH terms (manually or automatically assigned), UMLS
concepts (using MetaMap) or wikipedia, to modality filtration to more
complex language models that incorporate phrases (not just words) or paragraphs,
sentence selection and query translation, as well as techniques such as pseudo
relevance feedback. Many participants found the use of the manually assigned
MeSH terms to be most useful. Modality filtration, using either text–based or
image–based modality detection techniques was found to be useful by some
participants while others found only minimal benefit using the modality.</p>
        <p>Multimodal Retrieval This year, the run with the highest MAP utilized a
multimodal approach to retrieval. However, many groups that performed a pure
fusion of the text–based and image–based runs found a significant deterioration
in performance as the visual runs had very poor performance. This year’s results
again emphasize the previously noted observations that although the use of visual
information can improve the search results over purely textual methods, the
process of effectively combining the information from the captions and image
itself can be quite complex and are often not robust. Simple approaches of fusing
visual and textual runs rarely lead to optimized performance.
Run Name Retrieval Type Run Type Group MAP bPref P10
XRCE AX rerank comb.trec Mixed Automatic XRCE 0.3572 0.3841 0.4375
XRCE CHI2 LOGIT IMG MOD late.trec Mixed Automatic XRCE 0.3167 0.361 0.3812
XRCE AF LGD IMG late.trec Mixed Automatic XRCE 0.3119 0.3201 0.4375
WIKI AX IMG MOD late.trec Mixed Automatic XRCE 0.2818 0.3279 0.3875
OHSU all mh major all mod reorder.txt Mixed Automatic OHSU 0.256 0.2533 0.3813
OHSU high recall.txt Mixed Automatic OHSU 0.2386 0.2533 0.3625
queries terms 0.1 Modalities.trec Mixed Automatic ITI 0.1067 0.1376 0.2812
XRCE AX rerank.trec Mixed Automatic XRCE 0.0732 0.1025 0.1063
Exp Queries Cit CBIR CV MERGE MAXt Mixed Automatic ITI 0.0641 0.0962 0.1438
runMixt.txt Mixed Automatic UAIC2010 0.0623 0.0666 0.1313
Exp Queries Cit CBIR CAT MERGE MAX Mixed Automatic ITI 0.0616 0.0975 0.1375
Queries Citations CBIR CV MERGE MAX Mixed Automatic ITI 0.0583 0.0783 0.125
Multimodal-Rerank-ROI-QE-Merge Mixed Automatic ITI 0.0486 0.0803 0.1
NMFAsymmetricMixed k2 11 Mixed Automatic Bioingenium 0.0395 0.047 0.0438
GE Fusion img fulltext Vis0.2.run Mixed Automatic medGIFT 0.0245 0.0718 0.0375
GE Fusion img captions Vis0.2.run Mixed Automatic medGIFT 0.0208 0.0753 0.0375
Interactive Retrieval This year, as in previous years, interactive retrieval
was only used by a very small number of participants. The results were not
substantially better than automatic runs. This continues to be an area where we
would like to see improved participation but little success in doing so. For this
reason the manual and interactive runs are not shown in separate tables.
3.4</p>
      </sec>
      <sec id="sec-3-9">
        <title>Case–based Retrieval Results</title>
        <p>In terms of case–based retrieval almost all groups focused on using textual
retrieval techniques as combining visual retrieval on a case basis is a difficult
approach. Best results were obtained with a textual retrieval approach when
using relevance feedback.</p>
        <p>Visual Retrieval The performance of the single visual run submitted (see
Table 5) shows that the results are much lower than the text–based techniques. Still,
compared with the image–based retrieval only a single image–based run had a
higher MAP, meaning that also case–based retrieval is possible with purely visual
retrieval techniques and can be used as a complement to the text approaches.
Textual Retrieval The vast majority of submissions was in the category of
textual retrieval (see Table6). Best results were obtained by a collaboration of
IBM and UIUC in the textual part. Surprisingly the baseline text result of using
Lucene with the full text articles and with absolutely no optimization has the
third best result and is within the limit of statistical significance of the best run.
The first three runs are basically very close and then the performance slowly
drops of. In general results are slightly lower than for the image–based topics.
The baseline run using the image captions and then combining results of the
single images obtains a much lower performance.</p>
        <p>For the first time in several years there was actually a substantial number
of feedback runs, although only two groups submitted feedback runs (see
Table 7). These runs show that relevance feedback can improve results, although
the improvement is fairly low compared with the automatic run. All but one of
the feedback runs has very good results, showing that the techniques work in a
stable manner.</p>
        <p>Multimodal Retrieval Only two participants actually submitted a mixed
case–based result, and the performance of these runs is fairly low
highlighting the difficulty in combining the textual and visual results properly. Much
more research on the visual and combined retrieval seems necessary as the
current techniques in this field do not seem to work in a satisfying way. For this
reason an information fusion task using ImageCLEF 2009 data was organized
at ICPR 2010, showing an enormous increase in performance when good fusion
techniques are applied even when the base results have very strong variations in
performance [13]. Very few of these runs using more sophisticated fusion
techniques had a degradation in performance over the best single run.
3.5</p>
      </sec>
      <sec id="sec-3-10">
        <title>Relevance Judgement Analysis</title>
        <p>A number of topics, both image–based and case–based, were judged by two or
even three judges. Seven topics were judged by two judges while two additional
topics were judged by three judges. There were significant variations in the kappa
metric used to evaluate the inter–rater agreement. Kappa for these topics ranged
from 0 to 1. The average kappa was 0.47. However, there were 4 topics where
the kappa was zero as one judge had assessed no images as being relevant while
the other had said that 1–11 images were relevant. On the other hand, there was
a topic where both judges agreed that only a single image was relevant. Topics
with low number of relevant images (¡10) can cause difficulties in evaluation as
difference in opinions between judges one a single image can result in large
differences in performance metrics for that topic. Without these topics, the average
kappa was 0.657, a more acceptable figure.</p>
        <p>Run Retrieval Type Run Type Group MAP bPref P10
PhybaselineRelfbWMR 10 0.2sub Textual Feedback UIUCIBM 0.3059 0.3348 0.4571
PhybaselineRelfbWMD 25 0.2sub Textual Feedback UIUCIBM 0.2837 0.3127 0.4571
PhybaselineRelFbWMR 10 0.2 top20sub Textual Feedback UIUCIBM 0.2713 0.2897 0.4286
case based queries pico backoff 0.1.trec Textual Feedback ITI 0.1386 0.1666 0.2
PhybaselinefbWMR 10 0.2sub Textual Manual UIUCIBM 0.3551 0.3714 0.4714
PhybaselinefbWsub Textual Manual UIUCIBM 0.3441 0.348 0.4714
PhybaselinefbWMD 25 0.2sub Textual Manual UIUCIBM 0.3441 0.348 0.4714
case based expanded queries terms 0.1.trec Textual Manual ITI 0.0601 0.0825 0.0857</p>
        <p>Run Retrieval Type Run Type Group MAP bPref P10
case based queries cbir with case backoff Mixed Automatic ITI 0.0353 0.0509 0.0429
case based queries cbir without case backoff Mixed Automatic ITI 0.0308 0.0506 0.0214
GE Fusion case captions Vis0.2 Mixed Automatic medGIFT 0.0143 0.0657 0.0357
GE Fusion case fulltext Vis0.2 Mixed Automatic medGIFT 0.0115 0.0786 0.0357
We briefly explored the variability in the rankings of the various runs caused by
using different judges for a topic, especially on topics that had very few relevant
images. Topics 2 and 8 had a kappa of zero as one judge had not found any
relevant images in the pool with the other found 1 and 9 relevant images
respectively. Both judges had found one relevant image for topic 7. We explored the
changes in ranking caused by eliminating these topics from the evaluation. Most
runs had a none to substantial improvement in bpref with three runs
demonstrating a substantial improvement in rankings without these topics. However,
four runs had a drop in bpref as these runs had performed quite well on topic 7
and extremely well on topic 8. The relative rankings of the groups were vastly
unchanged with using the assessment of different judges aside from topics with
low number of relevant images.
4</p>
        <sec id="sec-3-10-1">
          <title>Conclusions</title>
          <p>As in 2009, the largest number of runs for the image–based and case–based
tasks used textual techniques. The semantic topics combined with a database
containing high–quality annotations lend themselves to textual methods.
However, unlike in 2009, the best runs were those that effectively combined visual
and textual methods. Visual runs continue to be rare and generally poor in
performance.</p>
          <p>Case–based topics had an increased participation over last year. As may be
expected based on the nature of the task, case–based retrieval is more easily
accomplished using textual techniques. Unlike in the ad–hoc runs, combining
visual image severely degraded the performance for case–based topics, meaning
that much more care needs to be taken with these combinations. More focus has
to be put on the combinations to increase performance. Maybe a pure fusion
task of results could be an additional challenge for the coming years.</p>
          <p>A kappa analysis between several relevance judgements for the same topics
shows that, although there are differences between judges, there was moderate
agreement on topics that have more than 10 relevant images. As a result topics
with very few relevant images could be removed or a more thorough testing could
already remove them during the topic creation process.</p>
          <p>For future campaign it seems important to explore how to effectively combine
visual techniques with the text–based methods. As has been stated at previous
ImageCLEFs, we strongly believe that interactive and manual retrieval are
important and we strive to improve participation in these. This year’s results show
that even simple feedback can significantly improve results.
5</p>
        </sec>
        <sec id="sec-3-10-2">
          <title>Acknowledgements</title>
          <p>We would like to thank the CLEF campaign for supporting the ImageCLEF
initiative. This work was partially funded by the Swiss National Science
Foundation (FNS) under contracts 205321–109304/1 and PBGE22–121204, the
American National Science Foundation (NSF) with grant ITR–0325160, Google, the
National Library of MEdicine grant K99LM009889, and the EU FP7 projects
Khresmoi and Promise. We would like to thank the RSNA for supplying the
images of their journals Radiology and Radiographics for the ImageCLEF
campaign.
9. Ojala, T., Pietikainen, M., Maenpaa, T.: Multiresolution gray–scale and rotation
invariant texture classification with local binary patterns. IEEE Transactions on
Pattern Analysis and Machine Intelligence 24(7) (July 2002) 971–987
10. Tamura, H., Mori, S., Yamawaki, T.: Texture features corresponding to visual
perception. IEEE Transactions on Systems, Man and Cybernetics 8(6) (1978)
460–472
11. Ma, W., Manjunath, B.: Texture features and learning similarity. In:
Proceedings of the 1996 IEEE Conference on Computer Vision and Pattern Recognition
(CVPR’96), San Francisco, California (June 1996) 425–430
12. Lowe, D.G.: Distinctive image features from scale-invariant keypoints.
International Journal of Computer Vision 60(2) (2004) 91–110
13. Mu¨ller, H., Kalpathy-Cramer, J.: The ImageCLEF medical retrieval task at icpr
2010 — information fusion to combine viusal and textual information. In:
Proceedings of the International Conference on Pattern Recognition (ICPR 2010). Lecture
Notes in Computer Science (LNCS), Istanbul, Turkey, Springer (August 2010) in
press.</p>
        </sec>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Clough</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          , Mu¨ller, H.,
          <string-name>
            <surname>Deselaers</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Grubinger</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lehmann</surname>
            ,
            <given-names>T.M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jensen</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hersh</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          :
          <article-title>The CLEF 2005 cross-language image retrieval track</article-title>
          .
          <source>In: Cross Language Evaluation Forum (CLEF</source>
          <year>2005</year>
          ). Springer Lecture Notes in Computer Science (
          <year>September 2006</year>
          )
          <fpage>535</fpage>
          -
          <lpage>557</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Clough</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          , Mu¨ller, H.,
          <string-name>
            <surname>Sanderson</surname>
            ,
            <given-names>M.:</given-names>
          </string-name>
          <article-title>The CLEF cross-language image retrieval track (ImageCLEF) 2004</article-title>
          . In Peters,
          <string-name>
            <given-names>C.</given-names>
            ,
            <surname>Clough</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            ,
            <surname>Gonzalo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            ,
            <surname>Jones</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.J.F.</given-names>
            ,
            <surname>Kluck</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            ,
            <surname>Magnini</surname>
          </string-name>
          , B., eds.:
          <article-title>Multilingual Information Access for Text, Speech and Images: Result of the fifth CLEF evaluation campaign</article-title>
          . Volume
          <volume>3491</volume>
          of Lecture Notes in Computer Science (LNCS).,
          <string-name>
            <surname>Bath</surname>
          </string-name>
          , UK, Springer (
          <year>2005</year>
          )
          <fpage>597</fpage>
          -
          <lpage>613</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3. Mu¨ller, H.,
          <string-name>
            <surname>Deselaers</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kim</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kalpathy-Cramer</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Deserno</surname>
            ,
            <given-names>T.M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Clough</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hersh</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          :
          <article-title>Overview of the ImageCLEFmed 2007 medical retrieval and annotation tasks</article-title>
          .
          <source>In: CLEF 2007 Proceedings. Volume 5152 of Lecture Notes in Computer Science (LNCS)</source>
          ., Budapest, Hungary, Springer (
          <year>2008</year>
          )
          <fpage>473</fpage>
          -
          <lpage>491</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Savoy</surname>
          </string-name>
          ,
          <source>J.: Report on CLEF-2001 experiments. In: Report on the CLEF Conference</source>
          <year>2001</year>
          (
          <article-title>Cross Language Evaluation Forum)</article-title>
          , Darmstadt, Germany, Springer LNCS 2406 (
          <year>2002</year>
          )
          <fpage>27</fpage>
          -
          <lpage>43</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5. Mu¨ller, H.,
          <string-name>
            <surname>Rosset</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          , Vall´ee,
          <string-name>
            <given-names>J.P.</given-names>
            ,
            <surname>Terrier</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            ,
            <surname>Geissbuhler</surname>
          </string-name>
          ,
          <string-name>
            <surname>A.</surname>
          </string-name>
          :
          <article-title>A reference data set for the evaluation of medical image retrieval systems</article-title>
          .
          <source>Computerized Medical Imaging and Graphics</source>
          <volume>28</volume>
          (
          <issue>6</issue>
          ) (
          <year>2004</year>
          )
          <fpage>295</fpage>
          -
          <lpage>305</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Kalpathy-Cramer</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hersh</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          :
          <article-title>Multimodal medical image retrieval: image categorization to improve search precision</article-title>
          .
          <source>In: MIR '10: Proceedings of the international conference on Multimedia information retrieval</source>
          , New York, NY, USA, ACM (
          <year>2010</year>
          )
          <fpage>165</fpage>
          -
          <lpage>174</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Radhouani</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hersh</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kalpathy-Cramer</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bedrick</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Understanding and improving image retrieval in medicine</article-title>
          .
          <source>Technical report</source>
          , Oregon Health and Science University (
          <year>2009</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Rosset</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          , Mu¨ller, H.,
          <string-name>
            <surname>Martins</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dfouni</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vall</surname>
            ´ee,
            <given-names>J.P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ratib</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          :
          <article-title>Casimage project - a digital teaching files authoring environment</article-title>
          .
          <source>Journal of Thoracic Imaging</source>
          <volume>19</volume>
          (
          <issue>2</issue>
          ) (
          <year>2004</year>
          )
          <fpage>1</fpage>
          -
          <lpage>6</lpage>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>