<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Extended Pixel Representation for Image Segmentation</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Deeptha</forename><surname>Girish</surname></persName>
							<email>girishde@mail.uc.edu</email>
							<affiliation key="aff0">
								<orgName type="department">EECS Department</orgName>
								<orgName type="institution">University of of Cincinnati Cincinnati</orgName>
								<address>
									<postCode>45221-0030</postCode>
									<region>OH</region>
									<country key="US">USA</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Vineeta</forename><surname>Singh</surname></persName>
							<email>singhvi@mail.uc.edu</email>
							<affiliation key="aff0">
								<orgName type="department">EECS Department</orgName>
								<orgName type="institution">University of of Cincinnati Cincinnati</orgName>
								<address>
									<postCode>45221-0030</postCode>
									<region>OH</region>
									<country key="US">USA</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Anca</forename><surname>Ralescu</surname></persName>
							<email>anca.ralescu@uc.edu</email>
							<affiliation key="aff0">
								<orgName type="department">EECS Department</orgName>
								<orgName type="institution">University of of Cincinnati Cincinnati</orgName>
								<address>
									<postCode>45221-0030</postCode>
									<region>OH</region>
									<country key="US">USA</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Extended Pixel Representation for Image Segmentation</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">4589F59EF57D6FB3A338596A3C61AA53</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-23T23:40+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>We explore the use of extended pixel representation for color based image segmentation using the K-means clustering algorithm. Various extended pixel representations have been implemented in this paper and their results have been compared. By extending the representation of pixels an image is mapped to a higher dimensional space. Unlike other approaches, where data is mapped into an implicit features space of higher dimension (kernel methods), in the approach considered here, the higher dimensions are defined explicitly. Preliminary experimental results which illustrate the proposed approach are promising.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Introduction</head><p>Image segmentation is a key processes in image analysis. This step is done irrespective of the goal of the analysis. It is one of the most critical tasks in this process. The aim of segmentation is to divide the image into non overlapping areas such that all pixels in one segment have similar features. This representation makes the image more meaningful and makes its analysis and understanding easier. For example, in medical images each color represents a particular stain that is essentially a protein binder that binds to a certain type of molecule. Therefore the different clusters obtained after segmenting such an image according to color can help us identify different biological components in that image. This can help us to do further analysis like finding meta data about each of the structures, finding one structures relative position to the other etc.</p><p>One way to achieve this is to cluster similar pixels into one cluster. Therefore, what defines similarity and how it is calculated becomes very important. A lot of research has been done on finding similarity measures for images and videos <ref type="bibr" target="#b2">(Wang et al.(2005)</ref>Wang, Zhang, and Feng), <ref type="bibr" target="#b1">(Huttenlocher et al.(1993)</ref>Huttenlocher, Klanderman, and Rucklidge), <ref type="bibr" target="#b0">(Gualtieri et al.(1992)</ref>Gualtieri, Le Moigne, and Packer). It is important that the measure we use captures the slightest difference in pixels. This plays an essential role in segmentation. <ref type="bibr">K-means (MacQueen et al.(1967)</ref>) is a popular algorithm that works well for image segmentation. The aim of the K-means algorithm as stated by Hartigan <ref type="bibr" target="#b0">ET. al. (Hartigan and Wong(1979)</ref>) is to divide M points in N dimensions into K clusters so that the within-cluster sum of squares distances is minimized. Thus, distance is an important factor. One of the major difficulties of image segmentation is to differentiate two similar regions which actually belong to different segments. This is because the existing features and the standard distances used to measure the dissimilarity do not capture the small differences very well. Increasing the sensitivity to small differences is the motivation for using the extended pixel representation. Image data is displayed on a computer as a bitmap which is a rectangular arrangement of pixels. The number of bits used to represent individual pixel has increased over the recent years, as computers have become more powerful. Todays computer systems often use a 24 bit color system. The most common color system in use is the RGB space (Red, Green and Blue).</p><p>Selecting the right color space for segmentation can be very important. Different parts of the image get highlighted better in different color spaces. It is also dependent on the purpose of the segmentation. To simplify this choice of color space, it is better to know in advance the features that represent maximum variation. This can be done by finding the coefficient of variation for each of the features of different color spaces. For a given data set, the coefficient of variation, defined as the ratio of standard deviation to the mean, represents the variability within the data set with respect to the mean. When selecting features/dimensions, it then makes sense to to take into consideration the coefficient of variation, more precisely, to select the features that have the highest variability. In selecting a color space, it is usual to adopt one (e.g., RGB, HSV) space. This study departs from this policy by selecting more than one color spaces, which further helps in better differentiate between pixels.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Related Work</head><p>The idea of extended pixel representation has been used previously to detect and repair halo in images <ref type="bibr" target="#b1">(Ohshima et al.(2006)</ref>Ohshima, Ralescu, Hirota, and Ralescu), to calculate the degree of difference between pixels in color images in order to identify the halo region. The extended pixel representation used in <ref type="bibr" target="#b1">(Ohshima et al.(2006)</ref>Ohshima, Ralescu, Hirota, and Ralescu) is </p><formula xml:id="formula_0">(R G B) (R G B RG RG GB GB BR BR ) (H S V ) (H S V HS HS SV SV V S V S ) (Y Cb Cr) (Y Cb Cr Y Cb Y Cb CbCr CbCr CrY CrY ) (R G B) (H S V ) (R G B H S V Y Cb Cr) (Y Cb Cr)</formula><p>described next. Given a pixel p in the image I, its extended pixel representation of p, is the real-valued vector</p><formula xml:id="formula_1">(R, G, B, RG , RG , GB , GB , BR , BR</formula><p>) where RGB are pixel values in the RGB color space, and XY and XY are the polar coordinates in the two dimensional color space, (X, Y ), with X, Y R, G, B, X = Y . This representation better support image segmentation because it captures the difference between pixels more accurately. In particular, it gives much better results for images with low contrast.</p><p>The same extended pixel representation can be implemented in other color spaces as well, as our experimental results for HSV , which uses H for Hue, synonymous to color, S for Saturation, the colorfulness of a color relative to its own brightness, and V for Value which refers to the lightness or darkness of a color. It is known that for most images, HSV color space is more suitable for segmentation and we show that the extended pixel representation for HSV improves the results even for this color space. The Y CbCr space uses Y for Luminance, the brightness of light, C for Chrominance, the color information of the signal, which is stored as two color-difference components (Cb and Cr). Table <ref type="table" target="#tab_0">1</ref> shows the features for each color space used in the experiments.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Current Approach</head><p>To illustrate the effect of the extended pixel representation, we consider two experiments as follows.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Sensitivity of distance to the extended space</head><p>Two pixels which are almost of the same color and having very similar R, G and B values are selected and their corresponding values in each of the extended pixel representations considered are calculated, using both the Euclidean distance and Manhattan distance. It is seen that the extended pixel representations are more sensitive to the dissimilarity between the pixels irrespective of the distance measure used. This is the idea behind using extended pixel representation. Especially for tasks like segmentation, this is a very efficient way to separate two samples which have similar feature values into different clusters.</p><p>Table <ref type="table" target="#tab_1">2</ref> shows the results for computing the distance between the two fixed pixels, in three standard color spaces, their extensions, and in an extended representation based on three color spaces.  <ref type="bibr">(10, 100, 150), p = (25, 120, 135), p = (30, 125, 130)</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Pixel Representation</head><p>(p 1 , p between pixels (in original color space and extended space) was used. For every image, the number of clusters was decided visually. The same number of clusters was used for all the extended pixel representations. The mean square error, equation ( <ref type="formula">1</ref>) and the signal to noise ratio, equation ( <ref type="formula">2</ref>) are calculated.</p><formula xml:id="formula_2">MSE(r, t) = 1 M ⇥ N M n=1 N m=1 [t(n, m) r(n, m)] 2 (1) Snr(t, r) = 10 log 10 M m=1 N n=1 [r(n, m)] 2 M m=1 N n=1 [r(n, m) t(n, m)] 2</formula><p>(2) where r(n, m) represents the original image and t(n, m) represents the segmented image, both of size [N, M ].</p><p>Table <ref type="table">5</ref> shows the results of the segmentation and MSE and Snr for each of these segmentations. Inspecting Table <ref type="table">5</ref> it is observed that for all images, even though the distance measure and the number of clusters were same, the mean square error was lower and correspondingly signal to noise ratio was higher when the extended pixel representation was used. It can be seen that the results look very similar to the original image because the clusters are colored with the mean color of all the pixels in that cluster. It can also be observed that the difference in results when the extended pixel representation is used is higher for images of low contrast. The extended pixel representation also performs better for pictures with high texture content.</p><p>The smaller changes in color values of the pixels are separated and recognized better in extended pixel representations. Irrespective of the color space used, the extended pixel representation gives lower mean squared error and higher signal to noise ratio than the standard pixel representation for all images. All pixels that are similar in color get clustered into one segment which might represent a particular structure in the image. Thus, this representation adds more meaning to the image.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Conclusion</head><p>The segmentation problem considered in this paper can be addressed by using the extended pixel representations.</p><p>The idea of representing each pixel with more information proved to be successful for image segmentation. It was observed that extended pixel representation can more effectively distinguish similar pixels. Although in the second experiment, it was seen that the percentage change of the euclidean distance between the extended pixel representation and the standard pixel representation when p was changed to p is very small, the effect turns out to be significant in the clustering step.With the same number of clusters and the same distance measure used, the mean square of the segmentation done using the extended pixel representation is lower for all images.</p><p>The promising experimental results of this idea encourages exploration of other extended pixel representations in the future. It will be interesting to include texture information like edges or frequency domain information as part of the pixel representation. Tailoring the extended pixel representations for the task in hand and automatic learning of the most important features and optimal number of features to represent a pixel for a particular task is an important idea to be considered for future work.  </p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Table 5 :</head><label>5</label><figDesc>Results on segmentation of six images, using each of the three color spaces separately (2nd row), extended pixel representation (3rd row), and all three color spaces (4th row). Bold font indicates the smallest MSE.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>`</head><label></label><figDesc></figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1 :</head><label>1</label><figDesc>The original and extended color spaces</figDesc><table><row><cell>Original</cell><cell>Extended</cell></row><row><cell>Color Space</cell><cell>Color Space</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2 :</head><label>2</label><figDesc>Euclidean distance (ED) and Manhattan distance (MD) for two fixed pixels in the original and extended color spaces, and the percentage of change due to the extended representation.</figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 4 :</head><label>4</label><figDesc>Changes in the Manhattan Distance (ED) when the pixel p changes to p . RGB pixel values are p =</figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head></head><label></label><figDesc>) (p 1 , p ) diff %</figDesc><table><row><cell>(R G B)</cell><cell>50</cell><cell>65</cell><cell>30</cell></row><row><cell>(R G B RG RG GB GB BR BR )</cell><cell cols="3">120.3 155.4 29.17</cell></row><row><cell>(H S V )</cell><cell cols="3">15.15 20.22 33.39</cell></row><row><cell>(H S V HS HS SV SV V S V S )</cell><cell cols="3">45.11 60.16 33.36</cell></row><row><cell>(Y Cb Cr)</cell><cell cols="3">27.40 35.82 30.73</cell></row><row><cell cols="4">(Y Cb Cr Y Cb Y Cb CbCr CbCr CrY CrY ) 82.13 106.1 29.22</cell></row><row><cell>(R G B H S V Y Cb Cr)</cell><cell cols="3">92.55 121.0 30.77</cell></row></table></figure>
		</body>
		<back>
			<div type="annex">
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Obviously, all distances in the extended space should be larger than those in the original space, because adding more features to the pixel representation in effects adds positive quantities to each of the distances considered and therefore this experiment merely confirms that theoretical fact.</p><p>Sensitivity of the distance in the original and extended space when the pixel changes</p><p>A second small experiment was run as follows. Starting with two pixels p, and p , ED and MD were computed just as in the previous section. Then p 2 was altered slightly, to obtain p , and the same distances were computed for p and p , and compared with those computed for p and p . Tables <ref type="table">3 and 4</ref> show the results.  <ref type="bibr">(10, 100, 150), p = (25, 120, 135), p = (30, 125, 130)</ref>. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Pixel Representation</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Experiments and Results</head><p>Segmentation using the k-means algorithm was implemented for three different images. Each image is in a different color space. Segmentation using K-means was done on these images using the standard pixel representation and the extended pixel representation. The Euclidean distance Deeptha Girish et al. <ref type="bibr">MAICS 2016 pp. 63-67</ref> </p></div>			</div>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Algorithm AS 136: A k-means clustering algorithm</title>
		<author>
			<persName><surname>Ja Gualtieri</surname></persName>
		</author>
		<author>
			<persName><surname>Le Moigne</surname></persName>
		</author>
		<author>
			<persName><forename type="first">;</forename><surname>Packer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>John</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Manchek</forename><forename type="middle">A</forename><surname>Hartigan</surname></persName>
		</author>
		<author>
			<persName><surname>Wong</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Frontiers of Massively Parallel Computation</title>
				<imprint>
			<publisher>IEEE</publisher>
			<date type="published" when="1979">1992. 1992. 1979</date>
			<biblScope unit="volume">28</biblScope>
			<biblScope unit="page" from="100" to="108" />
		</imprint>
	</monogr>
	<note>Distance between images</note>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Comparing images using the Hausdorff distance. Pattern Analysis and Machine Intelligence</title>
		<author>
			<persName><forename type="first">Gregory</forename><forename type="middle">A</forename><surname>Daniel P Huttenlocher</surname></persName>
		</author>
		<author>
			<persName><forename type="first">William J Rucklidge ;</forename><surname>Klanderman</surname></persName>
		</author>
		<author>
			<persName><surname>James Macqueen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the fifth Berkeley symposium on mathematical statistics and probability</title>
				<meeting>the fifth Berkeley symposium on mathematical statistics and probability<address><addrLine>Oakland, CA, USA</addrLine></address></meeting>
		<imprint>
			<date type="published" when="1967">1993. 1967. 2006</date>
			<biblScope unit="volume">15</biblScope>
			<biblScope unit="page" from="2" to="7" />
		</imprint>
	</monogr>
	<note>Proc. of the Conference IPMU</note>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">On the Euclidean distance of images. Pattern Analysis and Machine Intelligence</title>
		<author>
			<persName><forename type="first">Liwei</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Yan</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jufu</forename><surname>Feng</surname></persName>
		</author>
		<idno>YCbCr Snr=54.3042db MSE=0.00075</idno>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on</title>
		<imprint>
			<biblScope unit="volume">27</biblScope>
			<biblScope unit="issue">8</biblScope>
			<biblScope unit="page" from="1334" to="1339" />
			<date type="published" when="2005">2005</date>
		</imprint>
	</monogr>
	<note>d) Original Image e) Original Image f) Original Image d) RGB Snr=47.4465db MSE=0.0014 e</note>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">HSV HS HS SV SV VH VH Snr=47</title>
		<author>
			<persName><forename type="first">) Rgb Rg Rg Gb Gb</forename><surname>Br Br Snr=49</surname></persName>
		</author>
		<author>
			<persName><forename type="first">;</forename><surname>Mse=0 ; Mse=0</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">f) YCbCr YCb YCb CbCr CbCr CrY CrY Snr=54</title>
				<imprint>
			<publisher>MSE=0</publisher>
			<date type="published" when="2007">0012. 0044. 0007</date>
			<biblScope unit="page">9753</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">MSE=0.0013 e) RGB HSV YCbCr Snr=26</title>
	</analytic>
	<monogr>
		<title level="m">MSE=0.0038 e) RGB HSV YCbCr Snr=32</title>
				<imprint>
			<publisher>MSE=0</publisher>
			<date>0037</date>
			<biblScope unit="page">4912</biblScope>
		</imprint>
	</monogr>
	<note>d) RGB HSV YCbCr Snr=48</note>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
