<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Extended Pixel Representation for Image Segmentation</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Deeptha Girish</string-name>
          <email>girishde@mail.uc.edu</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Vineeta Singh</string-name>
          <email>singhvi@mail.uc.edu</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Anca Ralescu</string-name>
          <email>anca.ralescu@uc.edu</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>EECS Department University of of Cincinnati Cincinnati</institution>
          ,
          <addr-line>OH 45221-0030</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2016</year>
      </pub-date>
      <fpage>63</fpage>
      <lpage>67</lpage>
      <abstract>
        <p>We explore the use of extended pixel representation for color based image segmentation using the K-means clustering algorithm. Various extended pixel representations have been implemented in this paper and their results have been compared. By extending the representation of pixels an image is mapped to a higher dimensional space. Unlike other approaches, where data is mapped into an implicit features space of higher dimension (kernel methods), in the approach considered here, the higher dimensions are defined explicitly. Preliminary experimental results which illustrate the proposed approach are promising.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>Image segmentation is a key processes in image analysis.
This step is done irrespective of the goal of the analysis. It
is one of the most critical tasks in this process. The aim of
segmentation is to divide the image into non overlapping
areas such that all pixels in one segment have similar
features. This representation makes the image more
meaningful and makes its analysis and understanding easier.
For example, in medical images each color represents a
particular stain that is essentially a protein binder that
binds to a certain type of molecule. Therefore the
different clusters obtained after segmenting such an image
according to color can help us identify different biological
components in that image. This can help us to do further
analysis like finding meta data about each of the
structures, finding one structures relative position to the other etc.
One way to achieve this is to cluster similar pixels
into one cluster. Therefore, what defines similarity and how
it is calculated becomes very important. A lot of research
has been done on finding similarity measures for images
and videos (Wang et al.(2005)Wang, Zhang, and Feng),
(Huttenlocher et al.(1993)Huttenlocher, Klanderman, and
Rucklidge), (Gualtieri et al.(1992)Gualtieri, Le Moigne, and
Packer). It is important that the measure we use captures
the slightest difference in pixels. This plays an essential
role in segmentation. K-means (MacQueen et al.(1967)) is a
popular algorithm that works well for image segmentation.
The aim of the K-means algorithm as stated by Hartigan
ET. al. (Hartigan and Wong(1979)) is to divide M points
in N dimensions into K clusters so that the within-cluster
sum of squares distances is minimized. Thus, distance is
an important factor. One of the major difficulties of image
segmentation is to differentiate two similar regions which
actually belong to different segments. This is because the
existing features and the standard distances used to measure
the dissimilarity do not capture the small differences very
well. Increasing the sensitivity to small differences is the
motivation for using the extended pixel representation.
Image data is displayed on a computer as a bitmap which
is a rectangular arrangement of pixels. The number of bits
used to represent individual pixel has increased over the
recent years, as computers have become more powerful.
Todays computer systems often use a 24 bit color system.
The most common color system in use is the RGB space
(Red, Green and Blue).</p>
      <p>Selecting the right color space for segmentation can
be very important. Different parts of the image get
highlighted better in different color spaces. It is also dependent
on the purpose of the segmentation. To simplify this choice
of color space, it is better to know in advance the features
that represent maximum variation. This can be done by
finding the coefficient of variation for each of the features
of different color spaces. For a given data set, the coefficient
of variation, defined as the ratio of standard deviation to
the mean, represents the variability within the data set with
respect to the mean. When selecting features/dimensions, it
then makes sense to to take into consideration the coefficient
of variation, more precisely, to select the features that have
the highest variability. In selecting a color space, it is usual
to adopt one (e.g., RGB, HSV) space. This study departs
from this policy by selecting more than one color spaces,
which further helps in better differentiate between pixels.</p>
    </sec>
    <sec id="sec-2">
      <title>Related Work</title>
      <p>The idea of extended pixel representation has been used
previously to detect and repair halo in images (Ohshima
et al.(2006)Ohshima, Ralescu, Hirota, and Ralescu),
to calculate the degree of difference between pixels
in color images in order to identify the halo region.
The extended pixel representation used in (Ohshima
et al.(2006)Ohshima, Ralescu, Hirota, and Ralescu) is
described next. Given a pixel p in the image I, its
extended pixel representation of p, is the real-valued vector
(R, G, B, RG, RG, GB , GB , BR, BR) where RGB are
pixel values in the RGB color space, and XY and XY are
the polar coordinates in the two dimensional color space,
(X, Y ), with X, Y R, G, B, X = Y . This representation
better support image segmentation because it captures the
difference between pixels more accurately. In particular, it
gives much better results for images with low contrast.
The same extended pixel representation can be
implemented in other color spaces as well, as our experimental
results for HSV , which uses H for Hue, synonymous to
color, S for Saturation, the colorfulness of a color relative
to its own brightness, and V for Value which refers to the
lightness or darkness of a color. It is known that for most
images, HSV color space is more suitable for segmentation
and we show that the extended pixel representation for
HSV improves the results even for this color space. The
Y CbCr space uses Y for Luminance, the brightness of
light, C for Chrominance, the color information of the
signal, which is stored as two color-difference components
(Cb and Cr). Table 1 shows the features for each color space
used in the experiments.</p>
    </sec>
    <sec id="sec-3">
      <title>Current Approach</title>
      <p>To illustrate the effect of the extended pixel representation,
we consider two experiments as follows.</p>
    </sec>
    <sec id="sec-4">
      <title>Sensitivity of distance to the extended space</title>
      <p>Two pixels which are almost of the same color and having
very similar R, G and B values are selected and their
corresponding values in each of the extended pixel
representations considered are calculated, using both the Euclidean
distance and Manhattan distance. It is seen that the extended
pixel representations are more sensitive to the dissimilarity
between the pixels irrespective of the distance measure used.
This is the idea behind using extended pixel representation.
Especially for tasks like segmentation, this is a very efficient
way to separate two samples which have similar feature
values into different clusters.</p>
      <p>Table 2 shows the results for computing the distance
between the two fixed pixels, in three standard color spaces,
their extensions, and in an extended representation based on
three color spaces.</p>
      <p>HS HS SV SV V S V S )
(H S V )
(H S V
% of change
% of change
(Y Cb Cr) 29.15
(Y Cb Cr Y Cb Y Cb CbCr CbCr CrY CrY ) 48.24
(R G B H S V Y Cb Cr)
% of change with respect to average
distances for the original color spaces
Obviously, all distances in the extended space should
be larger than those in the original space, because adding
more features to the pixel representation in effects adds
positive quantities to each of the distances considered and
therefore this experiment merely confirms that theoretical
fact.</p>
    </sec>
    <sec id="sec-5">
      <title>Sensitivity of the distance in the original and extended space when the pixel changes</title>
      <p>A second small experiment was run as follows. Starting with
two pixels p, and p , ED and MD were computed just as in
the previous section. Then p2 was altered slightly, to obtain
p , and the same distances were computed for p and p , and
compared with those computed for p and p . Tables 3 and 4
show the results.
between pixels (in original color space and extended space)
was used. For every image, the number of clusters was
decided visually. The same number of clusters was used for all
the extended pixel representations. The mean square error,
equation (1) and the signal to noise ratio, equation (2) are
calculated.</p>
      <p>M SE(r, t) =</p>
      <p>1
M ⇥ N</p>
      <p>M</p>
      <p>N</p>
      <p>The smaller changes in color values of the pixels are
separated and recognized better in extended pixel
representations. Irrespective of the color space used, the
extended pixel representation gives lower mean squared
error and higher signal to noise ratio than the standard pixel
representation for all images. All pixels that are similar in
color get clustered into one segment which might represent
a particular structure in the image. Thus, this representation
adds more meaning to the image.</p>
    </sec>
    <sec id="sec-6">
      <title>Conclusion</title>
      <p>The segmentation problem considered in this paper can
be addressed by using the extended pixel representations.</p>
      <p>The idea of representing each pixel with more information
proved to be successful for image segmentation. It was
observed that extended pixel representation can more
effectively distinguish similar pixels. Although in the
second experiment, it was seen that the percentage change
of the euclidean distance between the extended pixel
representation and the standard pixel representation when
p was changed to p is very small, the effect turns out to
be significant in the clustering step.With the same number
of clusters and the same distance measure used, the mean
square of the segmentation done using the extended pixel
representation is lower for all images.</p>
      <p>The promising experimental results of this idea
encourages exploration of other extended pixel representations
in the future. It will be interesting to include texture
information like edges or frequency domain information
as part of the pixel representation. Tailoring the extended
pixel representations for the task in hand and automatic
learning of the most important features and optimal number
of features to represent a pixel for a particular task is an
important idea to be considered for future work.
a) RGB</p>
      <p>Snr=35.3260db
MSE=0.0050
b) HSV</p>
      <p>Snr=54.8959db
MSE = 0.0035</p>
      <sec id="sec-6-1">
        <title>c) YCbCr</title>
        <p>Snr=55.8133db
MSE=0.0031
a) RGB RG RG GB GB BR BR
b) HSV HS HS SV SV VH VH</p>
        <sec id="sec-6-1-1">
          <title>c) YCbCr YCb YCb CbCr CbCr CrY CrY</title>
          <p>Snr=39.3334db
MSE=0.0033
Snr=56.1241db
MSE=0.0026
Snr=58.8801db
MSE=0.0015
a)</p>
        </sec>
      </sec>
      <sec id="sec-6-2">
        <title>RGB HSV YCbCr</title>
        <p>Snr=37.4722db
MSE=0.0173</p>
      </sec>
      <sec id="sec-6-3">
        <title>b) RGB HSV YCbCr Snr=57.064db</title>
        <p>MSE=0.0021
c)</p>
      </sec>
      <sec id="sec-6-4">
        <title>RGB HSV YCbCr Snr=35.4811db MSE=0.0173</title>
      </sec>
      <sec id="sec-6-5">
        <title>d) Original Image</title>
      </sec>
      <sec id="sec-6-6">
        <title>e) Original Image</title>
      </sec>
      <sec id="sec-6-7">
        <title>f) Original Image</title>
        <p>d) RGB</p>
        <p>Snr=47.4465db
MSE=0.0014
e) HSV</p>
        <p>Snr=44.0748db
MSE=0.0062</p>
      </sec>
      <sec id="sec-6-8">
        <title>f) YCbCr</title>
        <p>Snr=54.3042db
MSE=0.00075
d) RGB RG RG GB GB BR BR
e) HSV HS HS SV SV VH VH</p>
        <sec id="sec-6-8-1">
          <title>f) YCbCr YCb YCb CbCr CbCr CrY CrY</title>
          <p>Snr=49.2535db
MSE=0.0012
Snr=47.3483db
MSE=0.0044
Snr=54.9753db
MSE=0.0007</p>
        </sec>
      </sec>
      <sec id="sec-6-9">
        <title>d) RGB HSV YCbCr Snr=48.1598db MSE=0.0013</title>
      </sec>
      <sec id="sec-6-10">
        <title>e) RGB HSV YCbCr Snr=26.9429db</title>
        <p>MSE=0.0038</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <string-name>
            <given-names>JA</given-names>
            <surname>Gualtieri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J Le</given-names>
            <surname>Moigne</surname>
          </string-name>
          , and
          <string-name>
            <given-names>CV</given-names>
            <surname>Packer</surname>
          </string-name>
          .
          <article-title>Distance between images</article-title>
          .
          <source>In Frontiers of Massively Parallel Computation</source>
          ,
          <year>1992</year>
          ., Fourth Symposium on the, pages
          <fpage>216</fpage>
          -
          <lpage>223</lpage>
          . IEEE,
          <year>1992</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          <string-name>
            <surname>John A Hartigan</surname>
          </string-name>
          and
          <article-title>Manchek A Wong. Algorithm AS 136: A k-means clustering algorithm</article-title>
          .
          <source>Journal of the Royal Statistical Society</source>
          . Series C (Applied Statistics),
          <volume>28</volume>
          (
          <issue>1</issue>
          ):
          <fpage>100</fpage>
          -
          <lpage>108</lpage>
          ,
          <year>1979</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          <string-name>
            <surname>Daniel P Huttenlocher</surname>
          </string-name>
          ,
          <article-title>Gregory A Klanderman,</article-title>
          and
          <string-name>
            <surname>William</surname>
          </string-name>
          J Rucklidge.
          <article-title>Comparing images using the Hausdorff distance</article-title>
          .
          <source>Pattern Analysis and Machine Intelligence</source>
          , IEEE Transactions on,
          <volume>15</volume>
          (
          <issue>9</issue>
          ):
          <fpage>850</fpage>
          -
          <lpage>863</lpage>
          ,
          <year>1993</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          <string-name>
            <surname>James</surname>
          </string-name>
          MacQueen et al.
          <article-title>Some methods for classification and analysis of multivariate observations</article-title>
          .
          <source>In Proceedings of the fifth Berkeley symposium on mathematical statistics and probability</source>
          , volume
          <volume>1</volume>
          , pages
          <fpage>281</fpage>
          -
          <lpage>297</lpage>
          . Oakland, CA, USA.,
          <year>1967</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          <string-name>
            <given-names>Chihiro</given-names>
            <surname>Ohshima</surname>
          </string-name>
          , Anca Ralescu, Kaoru Hirota, and
          <string-name>
            <given-names>Dan</given-names>
            <surname>Ralescu</surname>
          </string-name>
          .
          <article-title>Processing of halo region around a light source in color images of night scenes</article-title>
          .
          <source>In Proc. of the Conference IPMU</source>
          , volume
          <volume>6</volume>
          , pages
          <fpage>2</fpage>
          -
          <lpage>7</lpage>
          ,
          <year>2006</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          <string-name>
            <given-names>Liwei</given-names>
            <surname>Wang</surname>
          </string-name>
          , Yan Zhang, and
          <string-name>
            <given-names>Jufu</given-names>
            <surname>Feng</surname>
          </string-name>
          .
          <article-title>On the Euclidean distance of images</article-title>
          .
          <source>Pattern Analysis and Machine Intelligence</source>
          , IEEE Transactions on,
          <volume>27</volume>
          (
          <issue>8</issue>
          ):
          <fpage>1334</fpage>
          -
          <lpage>1339</lpage>
          ,
          <year>2005</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>