<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>IPL at CLEF 2012 Medical Retrieval Task</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Spyridon Stathopoulos</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Nikolaos Sakiotis</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Theodore Kalamboukis</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Informatics Athens University of Economics and Business</institution>
          ,
          <addr-line>Athens</addr-line>
          ,
          <country country="GR">Greece</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>This article presents an experimental evaluation on using Latent Semantic Analysis (LSA) for searching very large image databases. It also describes IPL's participation to the image CLEF ad-hoc textual and visual retrieval for the medical task in 2012. We report on our approaches and methods and present the results of our extensive experiments applying data fusion on the results obtained from LSA on several low-level visual features.</p>
      </abstract>
      <kwd-group>
        <kwd>LSA</kwd>
        <kwd>LSI</kwd>
        <kwd>CBIR</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>The continuous advances of internet and digital technologies, as well as the
rapidly increasing multimedia content used by modern information systems, have
imposed a need for an e cient system for organizing and retrieving content from
large multimedia collections. However, the performance in image retrieval is still
very far from being e ective for several reasons: computational cost, scalability
and performance.</p>
      <p>
        In our runs this year we have experimented with Latent Semantic Analysis
(LSA), a technique that, although has been used successfully in many
applications in the domain of text retrieval, [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] it has not experienced similar success
in CBIR. The main reason being the density of the f eatures images
matrix, C, generated in image retrieval, in contrast to textual retrieval where the
term document matrix is sparse. As a result, complexity cost of SVD is raised
to prohibitively high levels for both, space and computational time.
      </p>
      <p>In this article we give an overview of the application of our methods to ad-hoc
medical retrieval and present the results of our submitted runs. Our e orts this
year were concentrated on applying the LSA method to a number of low-level
visual features and then using data fusion techniques on the SVD transformed
low rank approximation of images to enhance retrieval. We explore a what we
call SVD-bypass technique to factor the feature matrix by solving a much smaller
in size eigenproblem of the term correlation matrix CCT instead of solving the
SVD of matrix C. This method proved to be a much more e cient and scalable
solution for large data sets.</p>
      <p>In the next section, we describe our approach and in the following sections
we present the submitted IPL's runs on textual and visual retrieval with their
corresponding descriptions and results. Finally, in the last section we conclude
the remarks of this work with propositions for future research.
2</p>
    </sec>
    <sec id="sec-2">
      <title>Visual Retrieval</title>
      <p>According to the traditional use of LSA in information retrieval a
term-bydocument matrix, C, is rst constructed and an SVD analysis is then performed
on this matrix. However as stated before, the feature matrix in the case of image
retrieval, is a dense matrix. This increases the computational costs of the SVD
analysis to prohibitive levels for large image databases. A typical example for
our database this year, for the color layout feature will produce a matrix of size
11288 305000 30GB (in double precision) which makes the SVD impossible
to solve with our computer resources. In our LSA implementation we solve the
eigenproblem of the feature correlation matrix CCT instead. This matrix, for
a suitable representation of the images is of a moderate size, demanding less
storage and the eigenvalue problem of CCT can be solved much faster than the
SVD factorization of the matrix C. We then approximate the feature matrix
taking only the k-largest eigenvalues and corresponding eigenvectors of matrix
C, for a suitable value of k.
2.1</p>
      <sec id="sec-2-1">
        <title>Preprocessing of the data</title>
        <p>It is well known that the representation of a digital image depends on several
factors, from its resolution to color models etc. In a collection of images, it is highly
possible that there will be important variations considering these characteristics.
Thus, each image undergoes through several transformations before the feature
extraction step. In our case we have applied the following transformations.
1. Size normalization. All images are re-scaled to the same size.
2. Transformation to gray-scale images.
3. Tile splitting. Each image is split into equal-sized, non-overlapping cells we
reference to as tiles.
2.2</p>
      </sec>
      <sec id="sec-2-2">
        <title>Feature Extraction and Selection</title>
        <p>
          The vector representation of the images was based on three low-level features of
MPEG-7, Scalable Color (SC) with 64 coe cients per tile, Color Layout (CL)
having 192 coe cients per tile and the Edge Histogram (EH) feature.
Experiments on CLEF 2011 image collection showed that, the extraction of the edge
histogram per tile had a negative impact on retrieval performance, thus this
feature was extracted from the whole image instead. All the features were extracted
using the Java library Caliph&amp;Emir of the Lire CBIR system [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ].
        </p>
        <p>Finally, a simple histogram with 32 levels of gray colors was extracted from
each tile. To increase the discriminating power of the histogram, we remove the
levels with high frequency and normalize the remaining histogram values for all
images. At the same time, all histogram levels with a total frequency above 80%
are considered stop-words and thus, they are removed. We refer to this feature
as Color Selection Histogram (CSH).
2.3</p>
        <p>Construction of the feature-correlation matrix CCT
As we have already mentioned the matrix C is full in the case of CBIR and so
it is the matrix CCT . This matrix multiplication is the most intensive part of
the computations and memory demanding. In our implementation we overcome
all these problems by splitting the matrix C into a number of blocks, such that
each block can be accommodated into the memory (C = (C1; C2; :::; Cp)) and
calculate CCT by:</p>
        <p>CCT =</p>
        <p>p
X CiCiT
i=1</p>
        <p>After solving the eigenproblem of the feature-correlation matrix CCT , the
k largest eigenvectors, say Uk, and the corresponding eigenvalues, are selected.
The original feature vectors are then projected into the k th dimensional space
using the transformation</p>
        <p>yk = UkT y
on the original vector representation of an image y.
2.4</p>
      </sec>
      <sec id="sec-2-3">
        <title>Data Fusion</title>
        <p>For the data fusion task we used a weighted linear combination of the results
obtained by using the LSA method on di erent features, as de ned by :
SCORE(Q; Image) =</p>
        <p>
          X wiscorei(Q; Image)
i
(3)
(1)
(2)
were scorei denotes the similarity score of an Image with respect to a feature i.
The weight of each feature type is determined as a function of its performance [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ].
The wi's were estimated by the square of the Mean Average Precision (MAP)
values attained by the corresponding feature on the CLEF '11 collection [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ].
These values are listed in Table 1.
        </p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Textual Retrieval</title>
      <p>
        This year's collection contains a subset of PubMed of 305000 images from PubMed.
A detailed description of the collection in given in the overview paper in [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. Since
records have the same structure as in previous CLEF collections over the past
years, we followed the same steps in the textual retrieval task as in CLEF 2011.
Each record is identi ed by a unique gureID, which is associated with: the title
of the corresponding article, the article URL, the caption of the gure, the pmid
and the gure URL. From the pmid we downloaded the MeSH terms assigned
to each article.
      </p>
      <p>Our retrieval system was based on the Lucene 1 search engine. For
indexing we removed stop-words and applied Porter's stemmer. For the multi- eld
retrieval the weights of the elds were assigned at indexing time. We kept the
same structure of the database as in CLEF 2009, 2010 and 2011. This year we
used only the default scoring function 2 which was best performing at the 2011
CLEF Ad-Hoc retrieval.</p>
      <p>
        Also we use the same weights for the elds as in the last three years [
        <xref ref-type="bibr" rid="ref6 ref7">6, 7</xref>
        ].
We used two sets of weights: one, that was estimated empirically on the CLEF
2009 collection and a second set where the weights were estimated by the value
of the Mean Average precision estimated on the CLEF 2010 collection.
1 http://lucene.apache.org/
2 http://lucene.apache.org/core/old versioned docs/versions/3 5 0/api/core/org/
apache/lucene/search/Similarity.html
      </p>
    </sec>
    <sec id="sec-4">
      <title>Experimental Results</title>
      <sec id="sec-4-1">
        <title>Results from Textual Retrieval</title>
        <p>This year we've submitted a total of six runs using di erent combinations of elds
and corresponding weights. In Table 2 we give the de nitions of our textual runs
and in Table 3 their corresponding results.
4.2</p>
      </sec>
      <sec id="sec-4-2">
        <title>Results from Visual Retrieval</title>
        <p>For the visual retrieval task, we've also submitted a total of six runs, using di
erent values for the parameter k which de nes the selected number of eigen-values
and vectors to use for indexing and retrieval. For all runs a data fusion on
various features was applied, using the weights acquired from runs of each individual
feature with the CLEF's 2011 collection. In Table 4 we give the de nitions of
our runs and in Table 5 their corresponding results.</p>
        <p>LSA with k=20 on 64 tiles
for Scalable Color, Color layout Color Selection Histogram
R5: IPL AUEB DataFusion LSA</p>
        <p>SC CL CSH 64seg 50k</p>
        <p>LSA with k=50 on 64 tiles
for Scalable Color, Color layout Color Selection Histogram
R6: IPL AUEB DataFusion LSA</p>
        <p>SC CL CSH 64seg 100k</p>
        <p>LSA with k=100 on 64 tiles
for Scalable Color, Color layout Color Selection Histogram</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Conclusions and Further work</title>
      <p>We have presented a new approach to LSA for CBIR replacing the SVD analysis
of the feature the matrix C (m n) by the solution of the eigenproblem for the
matrix CCT (m m). The method overcomes the high cost of SVD in terms
of memory and computing time. In addition, in all the experiments, of which
only a small part was submitted o cially this year, the optimal value of the
approximation parameter was less than 50 which makes the method attractive
for fusion with several low level features. Certainly our approach is promising
and has created new research directions that need further investigation. The
image representation has an impact on LSA performance and a more systematic
research on that direction is currently under progress. Also the eigenvalues of the
matrix CCT follow a Zip an distribution with the k-th largest values been well
separated giving small residual vectors to machine accuracy, which give us an
evidence on the stability of the calculated eigenvectors. More work is currently
underway in order to determine the stability of the proposed method.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Deerwester</surname>
            ,
            <given-names>S.C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dumais</surname>
            ,
            <given-names>S.T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Landauer</surname>
            ,
            <given-names>T.K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Furnas</surname>
            ,
            <given-names>G.W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Harshman</surname>
            ,
            <given-names>R.A.</given-names>
          </string-name>
          :
          <article-title>Indexing by latent semantic analysis</article-title>
          .
          <source>JASIS</source>
          <volume>41</volume>
          (
          <issue>6</issue>
          ) (
          <year>1990</year>
          )
          <volume>391</volume>
          {
          <fpage>407</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Lux</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chatzichristo s</surname>
          </string-name>
          , S.A.:
          <article-title>Lire: lucene image retrieval: an extensible java cbir library</article-title>
          . In
          <string-name>
            <surname>El-Saddik</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vuong</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Griwodz</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bimbo</surname>
            ,
            <given-names>A.D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Candan</surname>
            ,
            <given-names>K.S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jaimes</surname>
          </string-name>
          , A., eds.: ACM Multimedia, ACM (
          <year>2008</year>
          )
          <volume>1085</volume>
          {
          <fpage>1088</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Wu</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bi</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zeng</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Han</surname>
            ,
            <given-names>L</given-names>
          </string-name>
          .:
          <article-title>Assigning appropriate weights for the linear combination data fusion method in information retrieval</article-title>
          .
          <source>Inf. Process. Manage</source>
          .
          <volume>45</volume>
          (
          <year>July 2009</year>
          )
          <volume>413</volume>
          {
          <fpage>426</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Kalpathy-Cramer</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          , Muller, H.,
          <string-name>
            <surname>Bedrick</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Eggel</surname>
            , I., de Herrera,
            <given-names>A.G.S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tsikrika</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          :
          <article-title>Overview of the clef 2011 medical image classi cation and retrieval tasks</article-title>
          . In: CLEF (Notebook Papers/Labs/Workshop). (
          <year>2011</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5. Muller, H.,
          <string-name>
            <surname>de Herrera</surname>
            ,
            <given-names>A.G.S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kalpathy-Cramer</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fushman</surname>
            ,
            <given-names>D.D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Antani</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Eggel</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          :
          <article-title>Overview of the imageclef 2012 medical image retrieval and classi cation tasks</article-title>
          .
          <source>In: CLEF 2012 working notes</source>
          , Rome, Ital,
          <year>2012</year>
          (
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Gkoufas</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Morou</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kalamboukis</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          :
          <article-title>Ipl at imageclef 2011</article-title>
          . In: CLEF (Notebook Papers/LABs/Workshops). (
          <year>2011</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Gkoufas</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Morou</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kalamboukis</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          :
          <article-title>Combining textual and visual information for image retrieval in the medical domain</article-title>
          .
          <source>The Open Medical Informatics Journal</source>
          <volume>5</volume>
          (
          <year>2011</year>
          )
          <volume>50</volume>
          {
          <fpage>57</fpage>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>