<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Dimension Reduction Methods for Iris Recognition</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Pavel Moravec</string-name>
          <email>pavel.moravec@vsb.cz</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>V´aclav Sn´aˇsel</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Computer Science</institution>
          ,
          <addr-line>FEECS, VS</addr-line>
        </aff>
      </contrib-group>
      <fpage>80</fpage>
      <lpage>89</lpage>
      <abstract>
        <p>In this paper, we compare performance of several dimension reduction techniques, namely LSI, FastMap, and SDD in Iris recognition. We compare the quality of these methods from both the visual impact, and quality of generated "eigenirises". K. Richta, J. Pokorny´, V. Sn´aˇsel (Eds.): Dateso 2009, pp. 80-89, ISBN 978-80-01-04323-3.</p>
      </abstract>
      <kwd-group>
        <kwd>SVD</kwd>
        <kwd>FastMap</kwd>
        <kwd>information retrieval</kwd>
        <kwd>SDD</kwd>
        <kwd>iris recognition</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>1 Introduction
Methods of human identification using biometric features like fingerprint, hand
geometry, face, voice and iris are widely studied.</p>
      <p>
        A human eye iris has its unique structure given by pigmentation spots, furrows and
other tiny features which are stable throughout life. It is possible to scan an iris without
physical contact in spite of wearing contact lenses or eyeglasses. The iris is hard to forge
which makes the iris a suitable object for the identification of people. Iris recognition
seems to be more reliable than other biometric techniques like face recognition[
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. Iris
biometrics systems for both private and public use have been designed and deployed
commercially by NCR, Oki, IriScan, BT, US Sandia Labs, and others.
      </p>
      <p>
        In this paper, we use the Petland’s approach to image retrieval: image vectors of
complete images of the size width × height of the image [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] build the feature vectors.
      </p>
      <p>To fight the high dimension of image vector, we can extract several features which
represent the image and concatenate them into a feature vector. The feature extraction
methods can use different aspects of images as the features, typically the color features
(histograms), shape features (moments, contours, templates), texture features and
others (e.g. eigenvectors). Such methods are either using a heuristics based on the known
properties of the image collection, or are fully automatic and may use the original image
vectors as an input.</p>
      <p>In this paper we will concentrate on the last category – other feature extraction
methods which use known dimension reduction techniques and clustering for automatic
feature extraction.</p>
      <p>
        Singular value decomposition (SVD) was already successfully used for automatic
feature extraction. In case of face collection (such as our test data), the base vectors
can be interpreted as images, describing some common characteristics of several faces.
These base vectors are often called eigenfaces. For a detailed description of eigenfaces,
see [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ].
      </p>
      <p>However SVD is not suitable for huge collections and is computationally expensive,
so other methods of dimension reduction were proposed. We test two of them –
SemiDiscrete decomposition.</p>
      <p>
        Recently, Toeplitz matrix minimal Eigenvalues are also playing a role towards
image description and feature extraction [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]. This approach presents a method for the
reduction of image feature points as it deals with the geometric relation between the
points rather than their geometric position [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]. This can reduce the characteristic or
feature points number from n to at least 1n0 decreasing the computation level and hence
the task time.
      </p>
      <p>The rest of this paper is organized as follows. The second section explains used
dimension reduction methods. In the third section, we briefly describe qualitative
measures used for evaluation of our tests. In the next section, we supply results of tests
for several methods on ORL face collection. In conclusions we give ideas for future
research.
2</p>
      <p>Dimension Reduction Methods
We used three methods of dimension reduction for our comparison – Singular Value
Decomposition, Semi-Discrete Decomposition and FastMap, which are briefly described
bellow.
2.1</p>
      <p>
        Singular Value Decomposition
SVD [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] is an algebraic extension of classical vector model. It is similar to the Principal
components analysis (PCA) method, which was originally used for the generation of
eigenfaces. Informally, SVD discovers significant properties and represents the images
as linear combinations of the base vectors. Moreover, the base vectors are ordered
according to their significance for the reconstructed image, which allows us to consider
only the first k base vectors as important (the remaining ones are interpreted as “noise”
and discarded). Furthermore, SVD is often referred to as more successful in recall when
compared to querying whole image vectors [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ].
      </p>
      <p>Formally, we decompose the matrix of images A by singular value decomposition
(SVD), calculating singular values and singular vectors of A.</p>
      <p>We have matrix A, which is an n × m rank-r matrix and values σ1, . . . , σr are
calculated from eigenvalues (λi) of matrix AAT as σi = √λi. Based on them, we
can calculate column-orthonormal matrices U = (u1, . . . , ur) and V = (v1, . . . , vr),
where U T U = In a V T V = Im, and a diagonal matrix Σ = diag(σ1, . . . , σr), where
σi &gt; 0, σi ≥ σi+1.</p>
      <p>The decomposition</p>
      <p>A = U ΣV T
is called singular decomposition of matrix A and the numbers σ1, . . . , σr are singular
values of the matrix A. Columns of U (or V ) are called left (or right) singular vectors
of matrix A.</p>
      <p>Now we have a decomposition of the original matrix of images A. We get r nonzero
singular numbers, where r is the rank of the original matrix A. Because the singular
values usually fall quickly, we can take only k greatest singular values with the
corresponding singular vector coordinates and create a k-reduced singular decomposition of
A.</p>
      <p>Let us have k (0 &lt; k &lt; r) and singular value decomposition of A</p>
      <p>A = U ΣV T ≈ Ak = (UkU0)
Σk 0
0 Σ0</p>
      <p>VkT
V0T
We call Ak = UkΣkVkT a k-reduced singular value decomposition (rank-k SVD).</p>
      <p>Instead of the Ak matrix, a matrix of image vectors in reduced space Dk = ΣkVkT
is used in SVD as the representation of image collection. The image vectors (columns
in Dk) are now represented as points in k-dimensional space (the feature-space). For
an illustration of rank-k SVD see Figure 1.</p>
      <p>Rank-k SVD is the best rank-k approximation of the original matrix A. This means
that any other decomposition will increase the approximation error, calculated as a sum
of squares (Frobenius norm) of error matrix B = A−Ak. However, it does not implicate
that we could not obtain better precision and recall values with a different
approximation.</p>
      <p>To execute a query Q in the reduced space, we create a reduced query vector qk =
UkT q (another approach is to use a matrix Dk′ = VkT instead of Dk, and qk′ = Σk−1UkT q).
Instead of A against q, the matrix Dk against qk (or qk′) is evaluated.</p>
      <p>
        Once computed, SVD reflects only the decomposition of original matrix of images.
If several hundreds of images have to be added to existing decomposition (folding-in),
the decomposition may become inaccurate. Because the recalculation of SVD is
expensive, so it is impossible to recalculate SVD every time images are inserted. The
SVDUpdating [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] is a partial solution, but since the error slightly increases with inserted
images. If the updates happen frequently, the recalculation of SVD may be needed soon
or later.
Semidiscrete decomposition (SDD) is one of other LSI methods, proposed recently for
text retrieval in [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. As mentioned earlier, the rank-k SVD method (called truncated
SVD by authors of semidiscrete decomposition) produces dense matrices U and V , so
      </p>
      <p>Dk
Nonnegative
real values</p>
      <p>YkT</p>
      <p>Values {-1,0,1}
the resulting required storage may be even larger than the one needed by the original
term-by-document matrix A.</p>
      <p>To improve the required storage size and query time, the semidiscrete
decomposition was defined as</p>
      <p>A ≈ Ak = XkDkYkT ,
where each coordinate of Xk and Yk is constrained to have entries from the set ϕ =
{−1, 0, 1}, and the matrix Dk is a diagonal matrix with positive coordinates.</p>
      <p>The SDD does not reproduce A exactly, even if k = n, but it uses very little storage
with respect to the observed accuracy of the approximation. A rank-k SDD (although
from mathematical standpoint it is a sum on rank-1 matrices) requires the storage of
k(m + n) values from the set {−1, 0, 1} and k scalars. The scalars need to be only
single precision because the algorithm is self-correcting. The SDD approximation is
formed iteratively.</p>
      <p>The optimal choice of the triplets (xi, di, yi) for given k can be determined using
greedy algorithm, based on the residual Rk = A − Ak−1 (where A0 is a zero matrix).
2.3</p>
      <p>
        FastMap
FastMap [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] is a pivot-based technique of dimension reduction, suitable for Euclidean
spaces.
      </p>
      <p>In first step, it chooses two points (feature vectors) from the matrix A, which should
be most distant for calculated reduced dimension. Because it would be expensive to
calculate distances between all points, it uses following heuristics (all chosen points are
image vectors from matrix A):
1. A random point c0 is chosen.
2. The point bi having maximal distance δ(ci, bi) from ci is chosen, and based on it
we select the point ai with maximal distance δ(bi, ai)
3. We iteratively repeat step 2 with ci+1 = ai (authors suggest 5 iterations).
4. Points a = ai and b = bi in the last iteration are pivots for the next reduction step.</p>
      <p>In second step (having the two pivots a, b), we use the cosine law to calculate
position of each point on line joining a and b. The coordinate xi of point pi is calculated
as
xi =
δ2(ai, pi) + δ2(ai, bi) − δ2(bi, pi)
and the distance function for next reduction step is modified to</p>
      <p>δ′2(p′i, p′j ) = δ2(pi, pj ) − (xi − xj )2</p>
      <p>The pivots in original and reduced space are recorded and when we need to process a
query, it is projected using the second step of projection algorithm only. Once projected,
we can again use the original distance function in reduced space.
3</p>
    </sec>
    <sec id="sec-2">
      <title>Qualitative Measures of Retrieval Methods</title>
      <p>Since we need an universal evaluation of any retrieval method, we use some measures to
determine quality of such method. In case of Information Retrieval we usually use two
such measures - precision and recall. Both are calculated from the number of objects
relevant to the query Rel – determined by some other method, e.g. by manual annotation
of given collection and the number of retrieved objects Ret. Based on these numbers
we define precision (P ) as a fraction of retrieved relevant objects in all retrieved objects
and recall (R) as a fraction of retrieved relevant objects in all relevant objects. Formally:
P =
|Rel ∩ Ret|
|Ret|
and R =
|Rel ∩ Ret|
|Rel|</p>
      <p>
        So we can say that recall and precision denote, respectively, completeness of
retrieval and purity of retrieval. Unfortunately, it was observed that with the increase of
recall, the precision usually decreases [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. This means that when it is necessary to
retrieve more relevant objects, a higher percentage of irrelevant objects will be probably
obtained, too.
      </p>
      <p>
        For the overall comparison of precision and recall across different methods on a
given collection, we usually use the technique of rank lists [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], where we first sort the
distances from smallest to greatest and then go down through the list and calculate
maximal precision for recall closest to each of the 11 standard recall levels (0.0, 0.1,
0.2, . . . , 0.9, 1.0). If we are unable to calculate precision on i-th recall level, we take the
maximal precision for the recalls between i − 1-th and i + 1-th level. From all levels,
we calculate mean average, which is a single-value characteristics of overall
precisionrecall ratio.
4
      </p>
    </sec>
    <sec id="sec-3">
      <title>Experimental Results</title>
      <p>
        For testing of the different methods, we used iris collection consisting of 384 irises. The
iris were scanned by TOPCON optical device connected to the CCD Sony camera. The
acquired digitized image is RGB of size 576 × 768 pixels. Only the red (R) component
of the RGB image has been used in our experiments because it appears to be more
reliable than recognition based on green or blue components or converting the irises
to grayscale first. It is in accord with [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], where near-infrared wavelengths are used
anyway.
      </p>
      <p>We have excluded one of the three irises for each eye for further querying (so that
the query iris would not be included in the collection and skew the query results), which
led to a collection of 256 irises of 64 people.</p>
      <p>
        An example of several irises from the collection is shown in Figure 3, the first 12
query vectors are shown in Figure 8. We did not isolate the central part and eyelids to
provide comparable results with[
        <xref ref-type="bibr" rid="ref9">9</xref>
        ].
4.1
      </p>
      <p>Generated “Eigenirises” and Reconstructed Images
Many of tested methods were able to generate a set of base images, which could be
considered to be “eigenirises” as is the case of PCA, SVD and several other methods.
We are going to provide examples of both factors (base vectors) – “eigenirises” and
reconstructed images which can be obtained from regenerated Ak. We calculated results
for all methods in several dimensions, for the demonstration images we will use k = 64.
We do not provide these images for FastMap, where it is not possible (we could have
provided the images used as pivots in each step of FastMap process).</p>
      <p>With SVD, we obtain factors with different generality, the most general being among
the first. The first few are shown in figure 4. The eigenirises with higher index bring
more details to reconstructed images.</p>
      <p>The reconstructed images for rank-64 SVD method are somewhat blurred, but
generally still recognizable, which can bee observed in figure 5.</p>
      <p>The SDD method differs slightly from previous methods, since each factor contains
only values {−1, 0, 1}. Gray in the factors shown in figure 6 represents 0; −1 and 1 are
represented with black and white respectively.</p>
      <p>The images in figure 7 are reconstructed least exactly from all methods (although
consistently), but this is to be expected due to the three-valued encoding of base vectors.
One may note a general loss of fine details, which is unfortunate, since it means that the
query process would be highly affected and the retrieval results poor.
Fig. 6. First 18 base vectors (out of 100) for SDD method</p>
      <p>First, we calculated the mean average precision (MAP) for all relevant images in
rank lists. The relative MAPs (against original matrix A – 100%) are shown in Table 1.</p>
      <p>One would suspect, that querying in original dimension would provide better results
that any of the dimension reduction methods. In Table 2 we show the number of queries,
where the first returned iris was of the same person (out of 128).
5</p>
      <p>Conclusion
In this paper, we have compared several dimension reduction methods on real-live
image data (using L2 metrics). Whilst the SVD is known to provide quality results, it is
computationally expensive and in case we only need to beat the “curse of
dimensionality” by reducing the dimension, FastMap may suffice.</p>
      <p>
        There are some other newly-proposed methods, which may be interesting for
future testing, e.g. the SparseMap [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. Additionally, faster pivot selection technique for
FastMap may be considered. We may also benefit from the use of Toeplitz matrices and
their minimal eigenvalues relation.
What we currently need is a better iris segmentation, i.e. removing the central piece
(in our case in light reflection), eyelids, and identifying the exact iris position, such as
methods described in [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ].
      </p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <given-names>R.</given-names>
            <surname>Baeza-Yates</surname>
          </string-name>
          and
          <string-name>
            <given-names>B.</given-names>
            <surname>Ribeiro-Neto</surname>
          </string-name>
          .
          <article-title>Modern Information Retrieval</article-title>
          . Addison Wesley, New York,
          <year>1999</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <given-names>M.</given-names>
            <surname>Berry</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Dumais</surname>
          </string-name>
          , and
          <string-name>
            <given-names>T.</given-names>
            <surname>Letsche</surname>
          </string-name>
          .
          <article-title>Computational Methods for Intelligent Information Access</article-title>
          .
          <source>In Proceedings of the 1995 ACM/IEEE Supercomputing Conference</source>
          , San Diego, California, USA,
          <year>1995</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <given-names>J.</given-names>
            <surname>Daugman</surname>
          </string-name>
          .
          <article-title>Statistical richness of visual phase information: Update on recognizing persons by iris patterns</article-title>
          .
          <source>International Journal of Computer Vision</source>
          ,
          <volume>45</volume>
          (
          <issue>1</issue>
          ):
          <fpage>25</fpage>
          -
          <lpage>38</lpage>
          ,
          <year>2001</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <given-names>J.</given-names>
            <surname>Daugman</surname>
          </string-name>
          .
          <article-title>The importance of being random: statistical principles of iris recognition</article-title>
          .
          <source>Pattern Recognition</source>
          ,
          <volume>36</volume>
          (
          <issue>2</issue>
          ):
          <fpage>279</fpage>
          -
          <lpage>291</lpage>
          ,
          <year>2003</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <given-names>J.</given-names>
            <surname>Daugman</surname>
          </string-name>
          .
          <article-title>New methods in iris recognition</article-title>
          .
          <source>IEEE Transactions on Systems, Man, and Cybernetics</source>
          ,
          <string-name>
            <surname>Part</surname>
            <given-names>B</given-names>
          </string-name>
          ,
          <volume>37</volume>
          (
          <issue>5</issue>
          ):
          <fpage>1167</fpage>
          -
          <lpage>1175</lpage>
          ,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <given-names>C.</given-names>
            <surname>Faloutsos</surname>
          </string-name>
          and
          <string-name>
            <given-names>K.</given-names>
            <surname>Lin</surname>
          </string-name>
          .
          <article-title>FastMap: A Fast Algorithm for Indexing, Data-Mining and Visualization of Traditional and Multimedia Datasets</article-title>
          .
          <source>ACM SIGMOD Record</source>
          ,
          <volume>24</volume>
          (
          <issue>2</issue>
          ):
          <fpage>163</fpage>
          -
          <lpage>174</lpage>
          ,
          <year>1995</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <given-names>G. R.</given-names>
            <surname>Hjaltason</surname>
          </string-name>
          and
          <string-name>
            <given-names>H.</given-names>
            <surname>Samet</surname>
          </string-name>
          .
          <article-title>Properties of embedding methods for similarity searching in metric spaces</article-title>
          .
          <source>IEEE Transactions on Pattern Analysis and Machine Intelligence</source>
          ,
          <volume>25</volume>
          (
          <issue>5</issue>
          ):
          <fpage>530</fpage>
          -
          <lpage>549</lpage>
          ,
          <year>2003</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <given-names>T. G.</given-names>
            <surname>Kolda and D. P. O'Leary</surname>
          </string-name>
          .
          <article-title>Computation and uses of the semidiscrete matrix decomposition</article-title>
          .
          <source>In ACM Transactions on Information Processing</source>
          ,
          <year>2000</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <given-names>P.</given-names>
            <surname>Praks</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Machala</surname>
          </string-name>
          , and
          <string-name>
            <given-names>V.</given-names>
            <surname>Snášel</surname>
          </string-name>
          .
          <article-title>Iris Recognition Using the SVD-Free Latent Semantic Indexing</article-title>
          . In MDM/KDD2004 - Fifth International Workshop on
          <article-title>Multimedia Data Mining "Mining Integrated Media and Complex Data" in conjunction with KDD'2004 - The 10th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining; Section 2</article-title>
          .
          <source>Multimedia Data Mining: Techniques and Applications</source>
          , pages
          <fpage>67</fpage>
          -
          <lpage>71</lpage>
          , Seattle, WA, USA,
          <year>2004</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <given-names>K.</given-names>
            <surname>Saeed</surname>
          </string-name>
          .
          <article-title>Image Analysis for Object Recognition</article-title>
          . Bialystok Technical University Press, Bialystok, Poland,
          <year>2004</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <given-names>K.</given-names>
            <surname>Saeed and M. K. Nammous</surname>
          </string-name>
          .
          <article-title>A speech-and-speaker identification system: Feature extraction, description, and classification of speech-signal image</article-title>
          .
          <source>IEEE Trans. on Industrial Electronics</source>
          ,
          <volume>54</volume>
          (
          <issue>2</issue>
          ):
          <fpage>887</fpage>
          -
          <lpage>897</lpage>
          ,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12. G. Salton and
          <string-name>
            <given-names>G.</given-names>
            <surname>McGill</surname>
          </string-name>
          .
          <article-title>Introduction to Modern Information Retrieval</article-title>
          .
          <article-title>McGraw-ill</article-title>
          , New York, USA,
          <year>1983</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <given-names>M.</given-names>
            <surname>Turk</surname>
          </string-name>
          and
          <string-name>
            <given-names>A.</given-names>
            <surname>Pentland</surname>
          </string-name>
          .
          <article-title>Eigenfaces for recognition</article-title>
          .
          <source>Journal of Cognitive Neuroscience</source>
          ,
          <volume>3</volume>
          (
          <issue>1</issue>
          ):
          <fpage>71</fpage>
          -
          <lpage>86</lpage>
          ,
          <year>1991</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>