<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Robust Facial Feature Detection for Registration</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Hamid AMIRI</string-name>
          <email>hamidlamiri@yahoo.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>A. Geometry-based approaches</institution>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>LSTS Laboratory National School of Engineers of Tunis (ENIT)</institution>
          ,
          <country country="TN">Tunisia</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>-Face image analysis, in the context of computer vision, is, in general, about acquiring a high-level knowledge about a facial image. Facial feature extraction is important in many facerelated applications, such as face recognition, pose normalization, expression understanding and face tracking. Although there is not yet a general consensus, the fiduciary or first tier facial features, most often cited in the literature, are the eyes or the eye corners, the tip of the nose and the mouth corners. Similarly, second tier features are the eyebrows, the bridge of the nose, the tip of the chin, and so on. Registration based on feature point correspondence is one of the most popular methods for alignment. The importance of facial features stems from the fact that most face recognition algorithms in 2D and/or 3D rely on accurate feature localization. This step is critical not only directly for recognition techniques based on features themselves, but indirectly for the global appearance-based techniques that necessitate prior image normalization. For example, in 2D face recognition, popular recognition techniques, such as eigenfaces or Fisherfaces, are very sensitive to registration and scaling errors. For 3D face recognition, the widely used iterative closest point (ICP) registration technique requires scale-normalized faces and a fairly accurate initialization. In any case, both modalities require accurate and robust automatic landmarking. In this paper, we inspect shortcomings of existing approaches in the literature and propose a method for automatic landmarking of near-frontal faces. We show good detection results on different large image datasets under challenging imaging conditions.</p>
      </abstract>
      <kwd-group>
        <kwd>-component</kwd>
        <kwd>face Alignment</kwd>
        <kwd>feature extraction</kwd>
        <kwd>face registration</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>Taher KHADHRAOUI</p>
      <p>INTRODUCTION</p>
      <p>
        The face detection and facial feature extraction is
instrumental for the successful performance of subsequent
tasks in related computer vision applications. Many high-level
vision applications such as facial feature tracking, facial
modeling and animation, facial expression analysis, and face
recognition, require reliable feature extraction. Facial feature
points are referred to in the literature as "salient points",
"anchor points", or "facial landmarks" [
        <xref ref-type="bibr" rid="ref8">9</xref>
        ]. The most
frequently occurring fiduciary facial features are the four eye
corners, the tip of the nose, and the two mouth corners [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ].
Facial feature detection is a challenging computer vision
problem due to high inter-personal changes (gender, race), the
intra-personal variability (pose, expression) and acquisition
conditions (lighting, scale, facial accessories). To make valid,
more accurate, quantitative measurements in diverse
applications, it is needed to develop automated methods for
recognition. Generally, these systems include automatic
feature extractors and change trackers in the facial features in
static images.
      </p>
      <p>The paper is organized as follows. In Section 2 we review
some related work in the face image analysis. The proposed
approach is given in Section 3. The experimental results are
provided in Section 4. In Section 5, we discuss future work
and draw our conclusions.</p>
      <p>II.</p>
      <p>These methods extracted features using geometric
information such as relative positions and sizes of the face
components.</p>
      <p>Mauricio Hess and G. Martinez [2] used SUSAN algorithm to
extract facial features such as eye corners and center, mouth
corners and center, chin and cheek border, and nose corner etc.
Nevertheless these techniques require threshold, which, given
the prevailing sensitivity, may adversely affect the achieved
performance.</p>
      <p>This technique, matches the facial components to
previously designed templates using appropriate energy
functional. Genetic algorithms have been proposed for more
efficient searching times in template matching.</p>
      <p>
        The best match of a template in the facial image will yield the
minimum energy. Proposed by Yuille et al [
        <xref ref-type="bibr" rid="ref2">3</xref>
        ] these
algorithms require a priori template modeling, in addition to
their computational costs, which clearly affect their
performance.
      </p>
    </sec>
    <sec id="sec-2">
      <title>C. Colour segmentation techniques</title>
      <p>
        This approach makes use of skin color to isolate the face.
Any non-skin color region within the face is viewed as a
candidate for eyes and/or mouth. The performance of such
techniques on facial image databases is rather limited, due to
the diversity of ethnical backgrounds [
        <xref ref-type="bibr" rid="ref3">4</xref>
        ].
      </p>
      <p>
        These approaches generally use texture (intensity)
information only and learn the characteristics of the landmark
neighborhoods projected in a suitable subspace. Methods such
as principal component analysis [
        <xref ref-type="bibr" rid="ref4">5</xref>
        ], independent component
analysis [
        <xref ref-type="bibr" rid="ref5">6</xref>
        ], and Gabor wavelets [
        <xref ref-type="bibr" rid="ref6">7</xref>
        ] are used to extract the
feature vector.
      </p>
      <p>These approaches are commonly used for face recognition
rather than person identification.</p>
      <p>III.</p>
      <p>PROPOSED METHOD</p>
      <p>The proposed method is summarized in figure 4. It uses
four main steps:</p>
      <sec id="sec-2-1">
        <title>Face Detection</title>
      </sec>
      <sec id="sec-2-2">
        <title>Facial Feature Points in 2D</title>
      </sec>
      <sec id="sec-2-3">
        <title>Estimation of Face Pose</title>
      </sec>
      <sec id="sec-2-4">
        <title>Generation of 3D Geometric features</title>
        <p>Typical image analysis and in particular facial features
detection usually consists of several steps:</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>A. Face detection</title>
      <p>
        Face detection is a technology to determine the locations
and size of a human being face in a digital image. It only
detects facial expression and rest all in the image is treated as
background and is subtracted from the image. Region of
interest (ROI) is defined here as a facial feature candidate
point or region [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. Depending on the representation of the
facial features, methods can be divided to region based and
point based.
      </p>
      <p>In some applications preprocessing may increase the accuracy
of the localization. This applies mostly to the cases where the
acquisition parameters are insufficient, for example poor
lighting, noise or inadequate camera properties.</p>
      <p>Point based ROI detection can be performed in various ways.
Most of the facial features, for example eye corners, mouth
corners, nostrils, are placed on the edges.</p>
      <p>To build fully automated systems that analyze the information
contained in face images, robust and efficient face detection
algorithms are required. Given a single image, the goal of face
detection is to identify all image regions which contain a face
regardless of its three-dimensional position, orientation, and
lighting conditions. Such a problem is challenging because
faces are nonrigid and have a high degree of variability in size,
shape, color, and texture. These algorithms aim to find
structural features that exist even when the pose, viewpoint, or
lighting conditions vary, and then use the these to locate faces.</p>
    </sec>
    <sec id="sec-4">
      <title>B. Features extraction</title>
      <p>In the 2D processing part of the proposed method a number
of points are detected across the facial area of the current input
image.</p>
    </sec>
    <sec id="sec-5">
      <title>1) Nose holes.</title>
      <p>Finding nose holes in an area given from face's geometry
depends on the angle between camera and face. If there isn't a
direct line of sight between nose holes and camera, it is
obviously not possible to track them.</p>
      <p>Nose holes color have a significant saturation, depending on
its color black.</p>
      <p>The threshold must be defined and over geometry or clustering
two centers of saturation can be found.</p>
    </sec>
    <sec id="sec-6">
      <title>2) Mouth</title>
      <p>Detecting the middle of the mouth isn't as simple as it is
thought. There are a lot of possibilities, going over gradient
horizontal and/or vertical decent, hue or saturation. At the
moment it is implemented utilizing the distinct hue of lips.
Light reflects on lips and this point is fetched by a defined hue
value. In contrast to the methods, this method is not light
independent, thus intensity and direction of the light can
influence results. A better method should be included in the
future.</p>
    </sec>
    <sec id="sec-7">
      <title>3) Eyes and pupils</title>
      <p>A lot of ways can be developed to find pupils in the given
area surrounding the eyes, a Gabor eye-corner filter is
constructed to detect the corner point. It is more robust than
projection methods or edge detection methods.</p>
    </sec>
    <sec id="sec-8">
      <title>C. Estimation of Face Pose</title>
      <p>In order to correct those geometrically violated facial
features that deviate far from their actual positions, the
geometry constraint among the detected facial features is
imposed. However, in practice, the geometry variations among
all the thirty facial features under changes in individuals,
facial expressions and face orientations are too complicated to
be modelled.</p>
      <p>Estimation of face pose is a fundamental task in computer
vision. We infer face pose from geometric alignment of face
model and coarse face mesh reconstruction.</p>
      <p>Homogeneous model transformation matrix T4x4 formulates
the alignment (Eq. 1). The accurate triangular face mesh
model is gained from a previous range scan of the person
observed.</p>
      <p>We apply a least squares error metric E (Eq. 1) that minimizes
the sum of squared distances di from each point mi of model to
the plane containing the destination point Si, oriented
perpendicular to normal
T4×4: Homogeneous Model Transformation Matrix
mi, Si: Points of model and Reconstruction
Geometric alignment of mesh M3D (3) and point set W3D (4) is
realized with a variant of the Iterative Closest Point (ICP)
algorithm. In the ICP procedure we determine pose vector T
(5), which represents the optimal model transformation
parameters, with respect to an error metric.</p>
      <sec id="sec-8-1">
        <title>The ICP principle applied is as follows:</title>
        <p>

•</p>
        <p>Let cluster W3D (4) be a set of n points pi and M3D (3)
a surface model with m vertices aj and normals bj
Let CP( pi, aj ) be the closest vertex aj to a point pi
1. Let T[0] be an initial transformation estimate</p>
        <p>(5).</p>
        <p>2. Repeat for k = 1...kmax or until convergence:</p>
        <p>Compute the set of corresponding pairs S
• Compute the new transformation T[k] that minimizes</p>
        <p>Error metric E (2) w.r.t. all pairs S.</p>
        <p>Finally, we present a feature preserving Delaunay refinement
algorithm which can be used to generate 3D geometric feature.</p>
      </sec>
    </sec>
    <sec id="sec-9">
      <title>D. Generation of 3D Geometric Features</title>
      <p>3D structure of a face is estimated using the face feature
points. 3D measurements of any three points on a face can be
computed based on the perspective projection of a triangle.
These three feature points derived from the eyes and the
middle of the mouth</p>
      <p>P1(X1, Y1, Z1)
d3</p>
      <p>P2(X2, Y2, Z2)
d2</p>
      <p>d1</p>
      <p>P3(X3, Y3, Z3)
The 3D measurement of lengths d1, d2 and d3 of all edges of
the triangles are computed from the Equation (7).
(7)</p>
      <p>EXPERIMENTS RESULTS</p>
      <p>In our preliminary experiments, we have obtained some
promising results for different facial expressions (see Figure.
6). Thus, in order to determine a six degrees of freedom pose
vector, at least three points need to be found in the face, which
firstly, have to be visible also in a range of perspectives and
secondly, do not change during expression.</p>
      <p>Even more, these points need to be well distributed in space
and must be robustly and accurately detectable in the image.</p>
      <p>In this paper we have proposed an approach for the facial
feature detection which is efficient and fast to implement.
It provides a practical solution to the recognition problem. We
are currently investigating in more detail the issues of
robustness to changes in head size and orientation. Also we are
trying to recognize the gender of a person using the same
algorithm.
(a)
(b)
(c)
(d)</p>
      <p>Mauricio Hess and Geovanni Martinez, “Facial Feature Extraction based
on Smallest Univalue Assimilating Nucleus (SUSAN) Algorithm”.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>S.P.</given-names>
            <surname>Khandait</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.D.</given-names>
            <surname>Khandait</surname>
          </string-name>
          , and
          <string-name>
            <surname>Dr</surname>
          </string-name>
          .R.C.Thool, “
          <article-title>An Efficient Approach to Facial Feature Detection for Expression Recognition”</article-title>
          ,
          <source>International Journal of Recent Trends in Engineering</source>
          , Vol
          <volume>2</volume>
          , No. 1,
          <string-name>
            <surname>November</surname>
          </string-name>
          <year>2009</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>A.</given-names>
            <surname>Yuille</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Cohen</surname>
          </string-name>
          , and
          <string-name>
            <given-names>P.</given-names>
            <surname>Hallinan</surname>
          </string-name>
          , “
          <article-title>Facial feature extraction from faces using deformable templates”</article-title>
          ,
          <source>Proc. IEEE Computer Soc. Conf. On Computer Vision and Pattern Recognition</source>
          , pp.
          <fpage>104</fpage>
          -
          <lpage>109</lpage>
          ,
          <year>1989</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [4]
          <string-name>
            <surname>Yingjie</surname>
            <given-names>XIA</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kuang</surname>
            <given-names>MAO</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Wei</surname>
            <given-names>WANG</given-names>
          </string-name>
          , “
          <article-title>A Novel Facial Feature Localization and Extraction Method”</article-title>
          ,
          <source>Journal of Computational Information Systems</source>
          <volume>6</volume>
          :
          <issue>8</issue>
          (
          <year>2010</year>
          )
          <fpage>2779</fpage>
          -
          <lpage>2786</lpage>
          Available at http://www.Jofcis.com
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>K.</given-names>
            <surname>Susheel Kumar</surname>
          </string-name>
          , Shitala Prasad, Vijay Bhaskar Semwal, and
          <string-name>
            <given-names>R C</given-names>
            <surname>Tripathi</surname>
          </string-name>
          <article-title>, “Real Time Face Recognition Using Adaboost Improved Fast PCA Algorithm"</article-title>
          ,
          <source>International Journal of Artificial Intelligence &amp; Applications (IJAIA)</source>
          , Vol.
          <volume>2</volume>
          , No.3,
          <string-name>
            <surname>July</surname>
          </string-name>
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [6]
          <string-name>
            <surname>Antonini</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Popovici</surname>
          </string-name>
          and
          <string-name>
            <given-names>J. P.</given-names>
            <surname>Thiran</surname>
          </string-name>
          , ”
          <article-title>Independent Component Analysis and Support Vector Machine for Face Feature Extraction”</article-title>
          ,
          <source>Int. Conf. on Audio and Video-based Biometric Person Authentication</source>
          , pp.
          <fpage>111</fpage>
          -
          <lpage>118</lpage>
          , Guildford, UK,
          <year>2003</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [7]
          <string-name>
            <surname>Vukadinovic</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          and
          <string-name>
            <given-names>M.</given-names>
            <surname>Pantic</surname>
          </string-name>
          , ”
          <article-title>Fully Automatic Facial Feature Point Detection using Gabor Feature Based Boosted Classifiers”</article-title>
          ,
          <source>IEEE Int. Conf. on Systems, Man and Cybernetics</source>
          , Hawaii,
          <year>October 2005</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>Chai</given-names>
            <surname>Tong Yuen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Rizon</surname>
          </string-name>
          , Woo San San, and M.Sugisaka, “
          <article-title>Automatic Detection of Face and Facial Features”</article-title>
          ,
          <source>Proceedings of the 7th WSEAS International Conference on SIGNAL PROCESSING, ROBOTICS and AUTOMATION (ISPRA '08).</source>
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [9]
          <string-name>
            <surname>Sanqiang</surname>
            <given-names>ZHAO</given-names>
          </string-name>
          and
          <string-name>
            <surname>Yongsheng</surname>
            <given-names>GAO</given-names>
          </string-name>
          , “
          <article-title>Towards Robust and</article-title>
          Automated Head Pose Estimation: Elastic Energy Model”,
          <source>Biomedical Soft Computing and Human Sciences</source>
          , Vol.
          <volume>14</volume>
          , No.
          <issue>1</issue>
          , pp.
          <fpage>21</fpage>
          -
          <lpage>26</lpage>
          (
          <year>2009</year>
          ).
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>