<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Ethnicity Prediction Based on Iris Texture Features</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Stephen Lagree</string-name>
          <email>slagree@nd.edu</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Kevin W. Bowyer</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Computer Science and Engineering University of Notre Dame Notre Dame</institution>
          ,
          <addr-line>Indiana 46556</addr-line>
          <country country="US">USA</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>This paper examines the possibility of predicting ethnicity based on iris texture. This is possible if there are similarities of the iris texture of a certain ethnicity, and these similarities differ from ethnicity to ethnicity. This sort of “soft biometric” prediction could be used, for example, to narrow the search of an enrollment database for a match to probe sample. Using an iris image dataset representing 120 persons and 10-fold person-disjoint cross validation, we obtain 91% correct Asian / Caucasian ethnicity classification.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>
        Iris texture has been shown to be useful for biometric
identification and verification
        <xref ref-type="bibr" rid="ref1 ref2 ref5">(Bowyer, Hollingsworth,
and Flynn 2008; Phillips et al. 2005; Phillips et al. 2010;
Daugman 2006)</xref>
        . Studies have been done to determine if
iris texture contains information that can determine “soft
biometric” attributes of a person, such as ethnicity
        <xref ref-type="bibr" rid="ref10 ref11 ref2">(Qiu,
Sun, and Tan 2006; Qiu, Sun, and Tan 2007a)</xref>
        or gender
        <xref ref-type="bibr" rid="ref7">(Thomas et al. 2007)</xref>
        . This paper analyzes the possibility
of ethnicity prediction based on iris texture. The ability of
biometric systems to recognize the ethnicity of a subject
could allow automatic classification without human input.
Also, in an iris recognition system, an identification
request includes a “probe” iris, which is checked against a
“gallery” of enrolled images, to find the correct identity of
the requested iris. One application of this feature is to
narrow down the gallery of subjects to compare an iris to
for identification purposes. In a system with millions of
enrolled subjects, comparing an iris to all subjects could
take an extremely long time. Narrowing down the gallery
to only irises with the same ethnicity as the probe iris for
comparison could give a great speed improvement.
      </p>
    </sec>
    <sec id="sec-2">
      <title>Related Work</title>
      <p>
        The CASIA biometrics research group has performed
research on iris texture elements, including studies
        <xref ref-type="bibr" rid="ref10 ref10 ref11 ref11 ref2">(Qiu,
Sun, and Tan 2006; Qiu, Sun, and Tan 2007a; Qiu, Sun,
and Tan 2007b)</xref>
        on determining ethnicity based on iris
texture. To our knowledge, this is the only other work on
predicting ethnicity from iris texture. In
        <xref ref-type="bibr" rid="ref2">(Qiu, Sun, and
Tan 2006)</xref>
        , they report 86% accuracy in Asian / Caucasian
classification. Thomas et al. (2007) suggests that the work
in
        <xref ref-type="bibr" rid="ref2">(Qiu, Sun, and Tan 2006)</xref>
        may be biased due to
illumination differences in the two datasets the images
were taken from, the Asian subject images coming from
one dataset and the Caucasian subject images from another
dataset. If one dataset was generally brighter or darker
than the other, this factor could have entered into the
learned algorithm for separating the subjects based on
lighting, not iris texture. In the results presented in this
paper, we eliminate this issue by using images taken from
a single database to build our classifier, so that any
acquisition setup differences are just as likely to appear in
either ethnicity class. In
        <xref ref-type="bibr" rid="ref10 ref11">(Qiu, Sun, and Tan 2007a)</xref>
        , the
CASIA group reports 91% accuracy in Asian / non-Asian
ethnicity classification, using support vector machines and
texton features. The dataset in this work is composed of
2,400 images representing 60 different persons, so that
there are 20 images per iris. They divide the dataset into a
1,200-image training set and a 1,200-image test set, with
training and test set not specified to be person-disjoint. In
general, if iris images from the same person appear in both
the training and the test set, then the performance estimate
obtained is optimistically biased. In the results presented
in this paper, we eliminate this issue by using a
persondisjoint ten-fold cross-validation.
      </p>
      <p>
        In a study of how human observers categorize images,
Stark, Bowyer, and Siena (2010) found that humans
perceive general differences in iris texture that can be used
to classify iris textures into categories of similar texture
pattern. Observers grouped a set of 100 iris images into
categories of similar texture. The 100 images represented
100 different persons, and the 100 persons were balanced
on gender and on Asian / Caucasian ethnicity. The
observers did not know the gender or ethnicity of the
persons in the iris images. However, the grouping of
images into categories of similar iris texture resulted in
categories that were, on average, split 80% / 20% on
ethnicity. The same categories were on average divided
much more closely to 50% / 50% on gender. Thus, one
result of Stark’s work (2010) is that human observers
perceive consistent ethnicity-related differences in iris
texture. In this paper, we want to train a classifier to
explicitly perform the sort of ethnicity classification that
was found as a side effect of the texture similarity grouping
done by humans in
        <xref ref-type="bibr" rid="ref6">(Stark, Bowyer, and Siena 2010)</xref>
        and
that was previously explored in
        <xref ref-type="bibr" rid="ref10 ref11 ref2">(Qiu, Sun, and Tan 2006;
Qiu, Sun, and Tan 2007a)</xref>
        .
      </p>
    </sec>
    <sec id="sec-3">
      <title>Dataset</title>
      <p>
        We want to see how accurately we can identify ethnicity
based on iris texture. For this study we will use two
ethnicity classes, Caucasian and Asian. This study used
1200 iris images selected from the University of Notre
Dame’s iris image database. (This is a newer database than
was released to the iris biometrics research community for
the government’s Iris Challenge Evaluation (ICE) program
        <xref ref-type="bibr" rid="ref5">(Phillips et al. 2005; Phillips et al. 2010)</xref>
        .) All images
were obtained using an LG 4000 sensor at Notre Dame.
As with all commercial iris biometrics systems that we are
aware of, the images are obtained using near-infrared
illumination, and are 480x640 in size. One half of the
images, 600, were of persons whose ethnicity is classified
as Asian and the other half were from persons classified as
Caucasian. For each ethnicity, the 600 images represented
60 different persons, with 5 left iris images and 5 right iris
images per person. This 1,200-image dataset was
randomly divided into 10 folds of 120 images each, with 6
persons of each ethnicity in each fold. Thus the images in
the folds are person-disjoint; that is, each person’s images
appear in just one fold.
      </p>
    </sec>
    <sec id="sec-4">
      <title>Segmentation</title>
      <p>For this iris texture prediction study, we want to base our
findings solely on iris texture. Therefore we exclude
periocular clues that might be used as an indicator of
ethnicity. We segment the images to obtain the region of
interest, and mask out the eyelid-occluded portions of the
iris. We use Notre Dame’s IrisBee software to perform the
segmentations (Phillips et al. 2005). The output from
IrisBee that we use for texture examination is a 240x40
pixel normalized iris image along with the corresponding
bitmask of eyelid and eyelash occlusion locations. The
image segmentation and masking are exactly those that
would be used by IrisBEE in processing the images for
biometric recognition of a person’s identity. However, the
normalized images are not processed by the log-Gabor
filters that are used by IrisBEE to create the “iris code” for
biometric recognition. We create a different texture
feature vector for ethnicity prediction.</p>
    </sec>
    <sec id="sec-5">
      <title>Feature Generation</title>
      <p>
        After an image is segmented and normalized, we compute
texture features that can be used in training a classifier to
categorize images according to ethnicity. To do this we
apply different filters to the image at every non-masked
pixel location, and use the results of the filter to build a
feature vector. Six of the filters we have used are “spot
detectors” and “line detectors” of various sizes, as depicted
in Tables I to VI. For a given point in the image, if
applying a given filter would result in using any pixel that
is masked, then that filter application is skipped for that
point. The rest of the filters, depicted in Tables VI-VIII,
were created using Laws’ Texture Measures
        <xref ref-type="bibr" rid="ref3">(Laws 1980)</xref>
        .
These are designed to give responses for various types of
textures when convolved with images.
      </p>
      <p>A feature vector that describes the texture is computed
for each iris image. We divided the normalized image
array into a number of smaller sections in order to compute
statistics for sub-regions of the normalized image. This is
so that classification could be based on, for example,
relative differences between the band of the iris nearer the
pupil versus the band of the iris furthest from the pupil.
These regions were ten four-pixel horizontal bands and
four 60-pixel vertical bands of neighboring pixels in the
normalized iris image. The ten horizontal bands
correspond to concentric circular bands of the iris, running
from the pupil out to the sclera (white) of the eye. The
four vertical bands correspond roughly to the top, right,
bottom and left parts of the iris. Since the filters are
looking for different phenomena in the image, we find
statistics for the filter response of each image. Each image
contains 630 features, with 5 statistics calculated for each
of the 9 filters on all of the 14 regions. The five statistics
are: (1) average value of filter response, (2) standard
deviation of filter response, (3) 90th percentile value of
filter response, (4) 10th percentile value of filter response,
and (5) range between 90th and 10th percentile value. The
motivation for using the average value is to represent the
strength of a given spot size or line width in the texture.
The motivation for using the standard deviation is to
represent the degree of variation in the response. The
motivation for using the percentiles and range is to have an
alternate representation of the variation that is not affected
by small amounts of image segmentation error.
Results
We tried a variety of different classification algorithms
included in the WEKA package (Weka). This included
using meta-algorithms like Bagging with other classifiers.
By changing parameters, we achieved performance gains
on some of the algorithms. However, we found our best
results using the SMO algorithm with the default
parameters in WEKA for classification. The SMO
algorithm implements “Sequential Minimal Optimization”,
John Platt’s algorithm for building a support vector
machine classifier (Weka). The input to the SMO
algorithm is the feature vectors of all 1200 iris images that
we have computed. To assess the results of our classifier
we use cross-fold validation with ten folds using
stratification based on ethnicity. These folds are also
subject-disjoint to ensure the persons whose images are in
the test data have not been seen by the classification
algorithm in the training data.</p>
      <p>
        The SMO classifier results in higher accuracy compared
to a broad range of other classifiers, including decision tree
based algorithms and bagging. Using Bagging on the top
two classifiers, SMO and Random Forest, did not improve
performance. Running the experiment with the SMO
classifier and the feature vector as described above gives
us an accuracy of 90.58%. This is good accuracy,
representing an improvement on the 86% reported in
        <xref ref-type="bibr" rid="ref2">(Qiu,
Sun, and Tan 2006)</xref>
        and close to the 91% reported in
        <xref ref-type="bibr" rid="ref10 ref11">(Qiu,
Sun, and Tan 2007a)</xref>
        for a train-test split that was not
person-disjoint. When we do not use person disjoint
results, we see an accuracy of 96.17%, which is
significantly higher than Qiu, Sun, and Tan (2006; 2007a)
reported.
      </p>
      <p>We computed the classification accuracy for each
feature separately to see the impact of individual features.
Table X shows that some of the single features have almost
have the performance of all of the features together.
However none of them do as well as the combination of all
of the features. Some filters may be redundant; a
combination of a few might reproduce the performance of
all nine filters.</p>
      <p>To ensure that the size of our training dataset was not
limiting our accuracy levels, we ran the classifier with
different numbers of folds. Table XI shows the results we
achieved using 5, 10, and 20 fold cross validation. The
accuracy levels are all within one percent, indicating that
our performance should not be limited by our dataset size.</p>
    </sec>
    <sec id="sec-6">
      <title>Future Work</title>
      <p>To achieve even greater accuracy, we intend to implement
additional and more sophisticated features, and to look at
the effects of the size of the training set. We envision that
the number of different persons represented in the training
data is likely to be more important than the number of
images in the training set; that is, doubling the training set
by using twice as many images per person is likely not as
powerful as doubling the number of persons.</p>
      <p>For this experiment, we only looked at very broad
ethnicity classifications. More work could be done to
examine finer categories, such as Indian and Southeast
Asian. The performance of a classifier such as this has not
been tested on subjects of multiple ethnic backgrounds
either.</p>
    </sec>
    <sec id="sec-7">
      <title>Acknowledgments</title>
      <p>This work is supported by the Technical Support Working
Group under US Army contract W91CRB-08-C-0093, and
by the Central Intelligence Agency. The opinions, findings,
and conclusions or recommendations expressed in this
publication are those of the authors and do not necessarily
reflect the views of our sponsors.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <string-name>
            <surname>Bowyer</surname>
            ,
            <given-names>K.W.</given-names>
          </string-name>
          ; Hollingsworth,
          <string-name>
            <surname>K.</surname>
          </string-name>
          ; and Flynn,
          <string-name>
            <given-names>P. J.</given-names>
            <surname>Image</surname>
          </string-name>
          <article-title>Understanding for Iris Biometrics: A Survey, Computer Vision</article-title>
          and Image Understanding,
          <volume>110</volume>
          (
          <issue>2</issue>
          ),
          <fpage>281</fpage>
          -
          <lpage>307</lpage>
          , May
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          <string-name>
            <surname>Daugman</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <source>Probing the Uniqueness and Randomness of Iris Codes: Results From 200 Billion Iris Pair Comparisons, Proceedings of the IEEE, Nov</source>
          .
          <year>2006</year>
          ,
          <volume>94</volume>
          (
          <issue>11</issue>
          ),
          <fpage>1927</fpage>
          -
          <lpage>1935</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          <string-name>
            <surname>Laws</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Textured Image</surname>
            <given-names>Segmentation</given-names>
          </string-name>
          ,
          <source>Ph.D. Dissertation</source>
          , University of Southern California,
          <year>January 1980</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          <string-name>
            <surname>W.</surname>
          </string-name>
          <source>The Iris Challenge Evaluation</source>
          <year>2005</year>
          ,
          <article-title>Biometrics: Theory, Applications and Systems (BTAS 08)</article-title>
          ,
          <year>September 2008</year>
          , Washington, DC.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          <string-name>
            <surname>Phillips</surname>
            ,
            <given-names>P. J.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Scruggs</surname>
          </string-name>
          , W. T.;
          <string-name>
            <surname>O'Toole</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Flynn</surname>
            ,
            <given-names>P.J.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Bowyer</surname>
            ,
            <given-names>K.W.</given-names>
          </string-name>
          ; Schott,
          <string-name>
            <surname>C. L.</surname>
          </string-name>
          ; and Sharpe,
          <string-name>
            <surname>M.</surname>
          </string-name>
          <article-title>FRVT 2006</article-title>
          and
          <article-title>ICE 2006 Large-Scale Experimental Results</article-title>
          ,
          <source>IEEE Transactions on Pattern Analysis and Machine Intelligence</source>
          <volume>32</volume>
          (
          <issue>5</issue>
          ),
          <source>May</source>
          <year>2010</year>
          ,
          <fpage>831</fpage>
          -
          <lpage>846</lpage>
          ..
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          <string-name>
            <surname>Stark</surname>
            ,
            <given-names>L</given-names>
          </string-name>
          ; Bowyer,
          <string-name>
            <given-names>K.W.</given-names>
            ; and
            <surname>Siena</surname>
          </string-name>
          ,
          <string-name>
            <surname>S.</surname>
          </string-name>
          <article-title>Human categorization of iris texture patterns</article-title>
          ,
          <source>Biometrics Applications and Systems (BTAS)</source>
          ,
          <year>September 2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          <source>perceptual Theory</source>
          , Thomas,
          <string-name>
            <surname>V</surname>
          </string-name>
          ; Chawla,
          <string-name>
            <surname>N</surname>
          </string-name>
          ; Bowyer,
          <string-name>
            <surname>K. W.</surname>
          </string-name>
          ; and Flynn,
          <string-name>
            <surname>P. J.</surname>
          </string-name>
          <article-title>Learning to predict gender from iris images</article-title>
          .
          <source>In Proc. IEEE Int. Conf. on Biometrics: Theory</source>
          , Applications, and
          <string-name>
            <surname>Systems</surname>
          </string-name>
          , Sept
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          <string-name>
            <surname>Qiu</surname>
            ,
            <given-names>X. C.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Sun</surname>
            ,
            <given-names>Z. A.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Tan</surname>
            ,
            <given-names>T. N.</given-names>
          </string-name>
          <article-title>Global texture analysis of iris images for ethnic classification</article-title>
          .
          <source>In Springer LNCS 3832: Int.</source>
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          Conf. on Biometrics, pages
          <fpage>411</fpage>
          -
          <lpage>418</lpage>
          ,
          <year>June 2006</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          <string-name>
            <surname>Qiu</surname>
            ,
            <given-names>X. C.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Sun</surname>
            ,
            <given-names>Z. A.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Tan</surname>
            ,
            <given-names>T. N.</given-names>
          </string-name>
          <article-title>Learning appearance primitives of iris images for ethnic classification</article-title>
          .
          <source>In Int. Conf. on Image Processing</source>
          ,
          <string-name>
            <surname>pages</surname>
            <given-names>II</given-names>
          </string-name>
          :
          <fpage>405</fpage>
          -
          <lpage>408</lpage>
          ,
          <year>2007a</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          <string-name>
            <surname>Qiu</surname>
            ,
            <given-names>X. C.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Sun</surname>
            ,
            <given-names>Z. A.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Tan</surname>
            ,
            <given-names>T. N.</given-names>
          </string-name>
          <article-title>Coarse iris classification by learned visual dictionary</article-title>
          .
          <source>In Springer LNCS 4642: Int. Conf. on Biometrics</source>
          , pages
          <fpage>770</fpage>
          -
          <lpage>779</lpage>
          , Aug 2007b.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>