<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>ImageCLEF 2010 Modality Classification in Medical Image Retrieval: Multiple feature fusion with normalized kernel function</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Xian-Hua Han</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Yen-Wei Chen</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>College of Information Science and Engineering, Ritsumeikan University</institution>
          ,
          <addr-line>Kasatsu-shi, 525-8577</addr-line>
          ,
          <country country="JP">Japan</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>In this paper, we describe an approach for the automatic modality classification in medical image retrieval task of the 2010 CLEF cross-language image retrieval campaign (ImageCLEF). This work is focused on the process of feature extraction from medical images and fusion the different extracted visual feature and textual feature for modality classification. To extract visual features from the images, we used histogram descriptor of edge, gray or color intensity and block-based variation as global features and SIFT histogram as local feature, and the binary histogram of some predefined vocabulary words for image captions is used for textual feature. Then we combine the different features using normalized kernel functions for SVM classification. The proposed algorithm is evaluated by the provided modality dataset by ImageCLEF2010.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>Imaging modality is an important aspect of the image for medical retrieval [1-6]. In
user-studies, clinicians have indicated that modality is one of the most important filters
that they would like to be able to limit their search by. Many image retrieval websites
(Goldminer, Yottalook) allow users to limit the search results to a particular modality.
However, this modality is typically extracted from the caption and is often not correct
or present. Studies have shown that the modality can be extracted from the image itself
using visual features [7,8, 9]. Therefore, In this paper, we propose to use both visual and
textual features for medical image representation, and combine the different features
using normalized kernel function in SVM.</p>
      <p>In computer vision, studies have shown that the simple global features such as
histogram of edge, gray or color intensity and so on can represent images, and give the
acceptable performance in image retrieval or recognition research fields. Based on the
success of the above mentioned visual features for general image recognition, we also
use them as medical image representation for modality classification. Recently, using
local visual feature for image representation has been become very popular, and been
proved to be very effective for image categorization or retrieval [10]. The most famous
approach for image representation using local visual feature is bag of keypoints [11,
12]. The basic idea of bag of keypoints is that a set of local image patches is sampled
using some method (e.g. densely, randomly, or using a keypoint detector) and a vector
of visual descriptors is evaluated on each patch independently (e.g. SIFT descriptor,
normalized pixel values). The resulting distribution of descriptors in descriptor space is
then quantified in some way (e.g. by using vector quantization against a pre-specified
codebook to convert it to a histogram of votes for (i.e. patches assigned to) codebook
centres) and the resulting global descriptor vector is used as a characterization of the
image (e.g. as feature vector on which to learn an image classification rule based on an
SVM classifier). Furthermore, according to the visual properties of medical images, we
also calculate a histogram of small-block variance as visual feature for image
representation. For textual feature, we pre-define 90 vocabulary words somewhat according to
the statistical properties of training samples’ captions and our (not radiologist)
knowledge about medical modality, and calculate a binary histogram for any medical image
using their captions. After obtain the different feature for image representation, we
combine them together using kernel function for SVM classifier. Because different features
maybe have deferent scale and dimension, in order to allow each individual feature to
contribute equally for modality classification, we normalize the distance between two
samples using mean distance of all training samples, and then, obtain the kernel
function for each individual feature. The final kernel for SVM classification is the mean of
individual kernel, which can be called Joint Kernel Equal Contribution (JKEC). The
proposed algorithm is evaluated on the modality dataset of ImageCLEF2010, and the
classification rate is almost approximated the classification goal of the modality
classification task.
2</p>
    </sec>
    <sec id="sec-2">
      <title>Feature extraction for image representation</title>
      <p>In this section we describe how we extract a feature representation which is somewhat
robust to the high variability inherent in medical images and includes enough
discriminative information for modality category. As we known that it is difficult to classify
image categorization only with one type of image feature. So in this paper, we
represent images with differen images features: including gray and color intensity histogram,
block-based edge and variance histogram and popular bag-of-words model as visual
feature, and a binary histogram of the predefined vocabulary words for image captions
as textual feature. Then we merge them together for modality classification. Next, we
simply introduce the used features for medical image representation.
2.1</p>
      <sec id="sec-2-1">
        <title>Visual features</title>
        <p>
          Gray and Color intensity histogram: Intensity histograms are widely used to
capture the distribution information in an image. They are easy to compute and tend to
be robust against small changes of camera viewpoints. For Gray intensity histogram, we
can calculate the number of each intensity (
          <xref ref-type="bibr" rid="ref1 ref10 ref11 ref12 ref13 ref2 ref3 ref4 ref5 ref6 ref7 ref8 ref9">0–255</xref>
          ) for all image pixel, and normalize it
using pixel number. Given an image I in some color space (e.g., red, green, blue), for
calculate color histogram the color channels are quantized into a coarser space with k
bins for red, m bins for green and l bins for blue. Therefore the color histogram is a
vector h = (h1 ; h2 ; · · · ; hn )T , where n = k ml , and each element hi represents the
number of pixels of the discretized color in the image. We assume that all images have
been scaled to the same size. Otherwise, we normalize histogram elements as
(
          <xref ref-type="bibr" rid="ref1">1</xref>
          )
h′j =
        </p>
        <p>yj
∑n</p>
        <p>j=0 yj</p>
        <p>Block − based edge histogram: We firstly segment the image into several
blocks, and calculate edge histogram weighted by gradient intensity in each block. In
experiment, we grid-segment an image into 4 by 4 block, and calculate a 20-bin edge
histogram in each block. So we have 320 (20*16)-dimensional edge histogram feature
for medical image representation.</p>
        <p>Block − based variace histogram: For each pixel in an image, a small patch
centered by the specific pixel are used for calculating the local variation of the pixel.
after obtaining the local variation of all pixels in the image, a histogram of variation
intensity is calculated for the image representation.</p>
        <p>bag − of − words feature : In computer vision, local descriptors (i.e. features
computed over limited spatial support) have proved well-adapted to matching and
recognition tasks, as they are robust to partial visibility and clutter. In this paper, we use
gridsampling patches, and then compute appearance-based descriptors on the patches. In
contrast to the interest points from the detector, these points can also fall onto very
homogeneous areas of the image. After the patches are extracted,the SIFT [10] descriptor
is applied to represent the local features. The SIFT descriptor computes a gradient
orientation histogram within the support region. For each of 8 orientation planes, the gradient
image is sampled over a 4 by 4 grid of locations, thus resulting in a 128-dimensional
feature vector for each region. A Gaussian window function is used to assign a weight
to the magnitude of each sample point. This makes the descriptor less sensitive to small
changes in the position of the support region and puts more emphasis on the gradients
that are near the center of the region. To obtain robustness to illumination changes, the
descriptors are made invariant to illumination transformations of the form aI (x) + b by
scaling the norm of each descriptor to unity [10]. These SIFT features are then clustered
with a k-means algorithm using the Euclidean distance. Then we discard all information
for each patch except its corresponding closest cluster center identifier. For the test data,
this identifier is determined by evaluating the Euclidean distance to all cluster centers
for each patch. Thus, the clustering assigns a cluster c(x)21; :::C to each image patch
x and allows us to create histograms of cluster frequencies by counting how many of
the extracted patches belong to each of the clusters. The histogram representation h(X )
with C bins is then determined by counting and normalization such that:
hc(X ) =</p>
        <p>
          LX l=1
1 ∑LX (c; c(xl))
(
          <xref ref-type="bibr" rid="ref2">2</xref>
          )
where denotes the Kronecker delta function. Figure 2 shows the procedure
bag-ofwords(BoW) feature extraction and the extracted histogram feature of example images.
Obviously, there exist alternatives to algorithmic choices made in the proposed method.
For example, different interest point detectors can be used. However, it do not manifest
obvious merit for different background cluster of images. Furthermore, the
geometrical relation between the extracted patches is completely neglected in the approach
presented here. While this relation could be used to improve classification accuracy, it
remains difficult to achieve an effective reduction of the error rate in various situations
by doing so.
2.2
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>Textual features</title>
        <p>According to the statistical properties of word occurrence in each training modality
image’s captions and our prior knowledge about the classifying modality, we select 90
key-words, such as CT, Curve, MR, urethrogram, PET and so on, as the vocabulary for
forming a binary histogram for each medical image. The binary histogram for image
representation is 90-dimension vector, where each dimension is correspond to one
selected keyword. If one key-word is appeared one or more than one times in an image’s
caption, the value of the corresponding dimension in its represented binary histogram
will be 1, otherwise it will be 0.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3 feature fusion</title>
      <p>Given a training set (xi; yi)i=1;2; ;N of N instances consisting of an image xi ∈
and a class label yi ∈ 1; 2; · · · ; C, and given a set of F image features fm: →
ℜdm , m = 1; 2; · · · ; F , where dm denotes the dimensionality of the mth feature, the
problem of learning a classification function y : → 1; 2; · · · ; C from the features
and training set is called feature combination problem. In computer vision, the problem
of learning a multi-class classifier from training data is often addressed by means of
kernel methods. Kernel methods make use of kernel functions defining a measure of
similarity between pairs of instances. In the context of feature combination it is useful
to associate a kernel to each image feature as the following Eq. 3, and combine the
kernels of different features together. For a kernel function K of each feature between
real vectors we define the short-hand notation:</p>
      <p>
        Km(Ii; Ij ; ) = K(fm(Ii); fm(Ij )) = K(S(fm(Ii); fm(Ij )))
(
        <xref ref-type="bibr" rid="ref3">3</xref>
        )
where Ii and Ij are two samples, fm(Ii) is the mth extracted feature from the sample
Ii and S(fm(Ii); fm(Ij )) is the similarity measure between the mth features of the
samples Ii and Ij . Then the image kernel Km: × ∈ ℜ only considers similarity
with respect to image feature fm. If the image feature is specific to a certain aspect, say,
it only considers color information, then the kernel measures similarity only with regard
to this aspect. The subscript m of the kernel can then be understood as indexing into the
set of features. Because different features maybe have deferent scale and dimension, in
order to allow each individual feature to contribute equally for modality classification,
we normalize the distance between two samples using mean distance of all training
samples, and then, obtain the kernel function for each individual feature fm. The final
kernel for SVM classification is the mean of individual kernel, which can be called
Joint Kernel Equal Contribution (JKEC). For the feature similarity calculation of two
simples, we use 2 distance as the following:
L
∑ (xl − yl)2
S(fm(Ii); fm(Ij )) =
(
        <xref ref-type="bibr" rid="ref4">4</xref>
        )
where x and y represent the mth features fm(Ii); fm(Ij ) of samples i and j,
respectively, and xl is the lth element of the vector x. Then, the RBF function is used for
calculating the kernel:
      </p>
      <p>
        Km(Ii; Ij ; ) = exp( −S(fm(Ii); fm(Ij )) )
(
        <xref ref-type="bibr" rid="ref5">5</xref>
        )
where is the normalized item for Joint Equal Contribution of each feature. Here, we
use the distance mean of all training samples as , which will lead to similar
contribution of each feature to kernel. The proposed algorithm is evaluated on the modality
dataset of ImageCLEF2010, and the classification rate is almost approximated the
classification goal of the modality classification task.
4
      </p>
    </sec>
    <sec id="sec-4">
      <title>Experimental setup</title>
      <sec id="sec-4-1">
        <title>4.1 Image Data</title>
        <p>The database released for the ImageCLEF-2010 Medical modality classification in
medical retrieval task includes 2390 annotated modality images (CT: 314; GX: 355;
MR: 299; NM: 204; PET: 285; PX: 330; US: 307; XR:296) for training and a separate
evaluated set consisting of 2620 images. The aim is to automatically classify the
evaluated set using 8 different modality label sets including CT, MR, PET and so on. some
example images are shown in Fig. 1. A more detailed explanation of the database and
the tasks can be found in [13].
Fig. 2. The recognition rates of each modality category using different features: Edge, Gray, Color
and variance represent the recognition rates of Edge, Gray, Color and variance histogram,
respectively; BOF and textual mean the recognition rate of the BOF and textual features; Visual means
the recognition rates using combination of all visual features, and textual means those using only
textual feature; Visual+texual means those using combination of textual and visual features.
We evaluate classification performance of different features with SVM using training
dataset. From each modality set, 180 images are randomly selected for training, the
remainder are for test. We did this procedure 5 times, and gave the average classification
rate in Fig. 2 (a) for all modality category. From Fig. 2(a), It is obvious that different
features have different discriminant for each modality category. For an instance in all
visual features, the BOF (Bag-of-Feature) has the best recognition rate for MR
modality; however, the color histogram has the best result for PX modality. Therefore, after
combining all visual features together, the recognition rate for most of modality can be
greatly improved (Fig. 2(a)). The final results with combination of visual and textual
feature are also have large improvement that those with only visual or textual feature.
Figure. 2(b) gives the average classification rate for all modality. Based on the evaluated
better performance with visual feature than textual feature, we also use large weight for
visual feature when combine the visual and textual feature.
As Medical Image Processing Group (MIPG) of our Intelligent Image Processing
Laboratory (IIPL) In Ritsumeikan University, we prepared four runs for evaluation image
set, which used combine visual feature, textual feature, both visual and textual features
and weighted visual and textual features. The recognition results are shown in Table 1.
We submitted two runs using textual, combined textual and visual features by
on-linesystem, respectively.
5</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Conclusions</title>
      <p>In this paper, we proposed to extract different visual and textual features for medical
image representation, and fusion the different extracted visual feature and textual feature
for modality classification. To extract visual features from the images, we used
histogram descriptor of edge, gray or color intensity and block-based variation as global
features and SIFT histogram as local feature, and the binary histogram of some
predefined vocabulary words for image captions is used for textual feature. Because different
features maybe have deferent scale and dimension, in order to allow each individual
feature to contribute equally for modality classification, we proposed to use Joint
Kernel Equal Contribution (JKEC) for kernel fusion of different features. The proposed
algorithm is evaluated by the provided modality dataset by ImageCLEF2010.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Muller</surname>
            <given-names>H</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Michoux</surname>
            <given-names>N</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bandon</surname>
            <given-names>D</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Geissbuhler</surname>
            <given-names>A.</given-names>
          </string-name>
          ,
          <article-title>“ A review of content-based image retrieval systems in medicine clinical benefits and future directions”</article-title>
          ,
          <source>International Journal of Medical Informatics</source>
          <year>2004</year>
          ,
          <volume>73</volume>
          ,
          <fpage>1</fpage>
          -
          <lpage>23</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Muller</surname>
            <given-names>H</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Deselaers</surname>
            <given-names>T</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lehmann</surname>
            <given-names>T</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Clough</surname>
            <given-names>P</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hersh</surname>
            <given-names>W.</given-names>
          </string-name>
          , “
          <article-title>Overview of the ImageCLEFmed 2006 medical retrieval annotation tasks, Evaluation of Multilingual and Multimodal Information Retrieval”</article-title>
          ,
          <source>Seventh Workshop of the Cross-Language Evaluation Forum</source>
          , CLEF 2006, editors: Peters C,
          <string-name>
            <surname>Clough</surname>
            <given-names>P</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gey</surname>
            <given-names>F</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Karlgren</surname>
            <given-names>J</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Magnini</surname>
            <given-names>B</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Oard</surname>
            <given-names>DW</given-names>
          </string-name>
          ,
          <string-name>
            <surname>de Rijke</surname>
            <given-names>M</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Stempfhuber</surname>
            <given-names>M</given-names>
          </string-name>
          , LNCS
          <year>2006</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Hersh</surname>
            <given-names>W</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kalpathy-Cramer</surname>
            <given-names>J</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jensen</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <source>“ Medical Image Retrieval and Automated Annotation: OHSU at ImageCLEF</source>
          <year>2006</year>
          ”, Working Notes for the CLEF 2006 Workshop, Alicante, Spain.
          <year>2006</year>
          /
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Guld</surname>
            <given-names>MO</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kohnen</surname>
            <given-names>M</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Keysers</surname>
            <given-names>D</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Schubert</surname>
            <given-names>H</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wein</surname>
            <given-names>BB</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bredno</surname>
            <given-names>J</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lehmann</surname>
            <given-names>TM</given-names>
          </string-name>
          , “
          <article-title>Quality of DICOM header information for image categorization”</article-title>
          ,
          <string-name>
            <surname>Proceedings</surname>
            <given-names>SPIE</given-names>
          </string-name>
          ,
          <volume>4685</volume>
          ,
          <fpage>280</fpage>
          -
          <lpage>287</lpage>
          ,
          <year>2002</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Hersh</surname>
            <given-names>W</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Muller</surname>
            <given-names>H</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jensen</surname>
            <given-names>J</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yang</surname>
            <given-names>J</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gorman</surname>
            <given-names>P</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ruch</surname>
            <given-names>P.</given-names>
          </string-name>
          , “
          <article-title>Advancing biomedical image retrieval: development and analysis of a test collection”</article-title>
          ,
          <source>J Amer Med Inform Assoc</source>
          <year>2006</year>
          ;
          <volume>13</volume>
          (
          <issue>5</issue>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Kalpathy-Cramer</surname>
            <given-names>J</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hersh</surname>
            <given-names>W.</given-names>
          </string-name>
          , “
          <article-title>Automatic image modality based classification and annotation to improve medical image retrieval”</article-title>
          ,
          <source>Stud Health Technol Inform</source>
          .
          <year>2007</year>
          ;
          <volume>129</volume>
          (Pt 2):
          <fpage>1334</fpage>
          -
          <lpage>8</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <given-names>Alex</given-names>
            <surname>Pentland</surname>
          </string-name>
          , Rosalind Picard, and Stan Sclaro, “Photobook:
          <article-title>Content-based manipulation of image databases”</article-title>
          ,
          <source>International Journal of Computer Vision</source>
          ,
          <volume>18</volume>
          (
          <issue>3</issue>
          ):
          <fpage>233</fpage>
          -
          <lpage>254</lpage>
          ,
          <year>June 1996</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <given-names>Abolfazl</given-names>
            <surname>Lakdashti</surname>
          </string-name>
          and
          <string-name>
            <surname>M.</surname>
          </string-name>
          <article-title>Shahram Moin, “A New Content-Based Image Retrieval Approach Based on Pattern Orientation Histogram”</article-title>
          , Vol. ume 4418,
          <string-name>
            <surname>Computer</surname>
            <given-names>Vision</given-names>
          </string-name>
          /Computer Graphics Collaboration Techniques,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <given-names>A.K.</given-names>
            <surname>Jain</surname>
          </string-name>
          and
          <string-name>
            <given-names>A.</given-names>
            <surname>Vailaya</surname>
          </string-name>
          , “
          <article-title>Image retrieval using color and shape”</article-title>
          ,
          <source>Pattern Recognition</source>
          , Vol.
          <volume>29</volume>
          , No.
          <issue>8</issue>
          , pp.
          <fpage>1233</fpage>
          -
          <lpage>1244</lpage>
          ,
          <year>1996</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10. D. Lowe, “
          <article-title>Distinctive image features from scale-invariant keypoints”</article-title>
          ,
          <source>International Journal of Computer Vision</source>
          ,
          <volume>60</volume>
          (
          <issue>2</issue>
          ),pp.
          <fpage>91</fpage>
          -
          <lpage>110</lpage>
          ,
          <year>2004</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11. G. Csurka,
          <string-name>
            <given-names>C.</given-names>
            <surname>Dance</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Fan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Willamowski</surname>
          </string-name>
          , and
          <string-name>
            <given-names>C.</given-names>
            <surname>Bray</surname>
          </string-name>
          , “
          <article-title>Visual categoraization with bags of keypoints”</article-title>
          ,
          <source>In Proc. ECCVWorkshop on Statistical Learning in Computer Vision</source>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>16</lpage>
          ,
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Lazebnik</surname>
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Schmid</surname>
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ponce</surname>
            <given-names>J</given-names>
          </string-name>
          , ”
          <article-title>Beyond Bags of Features: Spatial Pyramid Matching for Recognizing Natural Scene Categories”</article-title>
          ,
          <source>In Proc. CVPR</source>
          , pp.
          <fpage>2169</fpage>
          -
          <lpage>2178</lpage>
          ,
          <year>2006</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Henning</surname>
            <given-names>Muller</given-names>
          </string-name>
          , Jayashree Kalpathy-Cramer, Ivan Eggel, Steven Bedrick,
          <string-name>
            <given-names>Charles E. Kahn</given-names>
            <surname>Jr.</surname>
          </string-name>
          , and William Hersh, “
          <article-title>Overview of the CLEF 2010 medical image retrieval track”</article-title>
          ,
          <source>In the Working Notes of CLEF</source>
          <year>2010</year>
          , Padova, Italy,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>