<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">KIDS&apos;s evaluation in medical image retrieval task at ImageCLEF 2004</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Pei-Cheng</forename><surname>Cheng</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Department of Computer &amp; Information Science</orgName>
								<orgName type="institution">National Chiao Tung University</orgName>
								<address>
									<addrLine>1001 Ta Hsueh Rd</addrLine>
									<postCode>30050</postCode>
									<settlement>Hsinchu</settlement>
									<region>O.C</region>
									<country>TAIWAN, R</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Been-Chian</forename><surname>Chien</surname></persName>
							<email>bcchien@ipx.ntntc.edu.tw</email>
						</author>
						<author>
							<persName><forename type="first">Hao-Ren</forename><surname>Ke</surname></persName>
							<affiliation key="aff2">
								<orgName type="institution" key="instit1">University Library</orgName>
								<orgName type="institution" key="instit2">National Chiao Tung University</orgName>
								<address>
									<addrLine>1001 Ta Hsueh Rd</addrLine>
									<postCode>30050</postCode>
									<settlement>Hsinchu</settlement>
									<region>O.C</region>
									<country>TAIWAN, R</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Wei-Pang</forename><surname>Yang</surname></persName>
							<email>wpyang@cis.nctu.edu.tw</email>
						</author>
						<author>
							<affiliation key="aff1">
								<orgName type="department">Department of Information Engineering</orgName>
								<orgName type="institution">I-Shou University 1</orgName>
								<address>
									<addrLine>Section 1, Hsueh-Cheng Rd., Ta-Hsu Hsiang</addrLine>
									<postCode>840</postCode>
									<settlement>Kaohsiung</settlement>
									<region>O.C</region>
									<country>TAIWAN, R</country>
								</address>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff3">
								<orgName type="department">Department of Information Management</orgName>
								<orgName type="institution">National Dong-Hwa University</orgName>
								<address>
									<addrLine>1, Sec. 2, Da Hsueh Rd., Shou-Feng</addrLine>
									<settlement>Hualien</settlement>
									<country>Taiwan, R.O.C</country>
								</address>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff4">
								<orgName type="department">Department of Computer Science and Information Engineering</orgName>
								<orgName type="institution">National University of TAINAN</orgName>
								<address>
									<addrLine>33, Sec. 2, Su-Lin Street</addrLine>
									<postCode>700</postCode>
									<settlement>Tainan</settlement>
									<country>Taiwan, R.O</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">KIDS&apos;s evaluation in medical image retrieval task at ImageCLEF 2004</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">9A846A3F5103AB27C497CFF0A592D5A3</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T17:50+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Medical image retrieval</term>
					<term>Color histogram</term>
					<term>Relevance feedback</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>We describe our participation in the medical retrieval task of the ImageCLEF 2004. This task aims at finding images that are similar with respect to modality (CT, radiograph, MRI, and so on). We propose several image features, including color histogram, gray-spatial histogram, coherence moment, and gray correlogram, to facilitate the retrieval of similar images. The initial retrieval results are obtained via visual feature analysis. An automatic feedback mechanism clusters visually and textually similar images among these initial results to help refine the query. In this paper, we present the system used, focusing on novel and newly developed aspects. The evaluated result shows that the automatic feedback mechanism improves the precision by 15%.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>The importance of digital image retrieval techniques increases in the emerging fields of medical imaging and picture archiving and communication systems (PACS). The increasing reliance of modern medicine on diagnostic techniques such as radiology, histopathology, and computerized tomography has led to the explosion in the number and importance of medical images now stored by most hospitals. While the prime requirement for medical imaging systems is to be able to display images relating to a named patient, there is increasing interest in the use of CBIR (content-based image retrieval) techniques to aid diagnosis by identifying similar past cases.</p><p>In the past years, content-based image retrieval has been one of the most hot research areas in the field of computer vision. The commercial QBIC <ref type="bibr" target="#b2">[Flickner95]</ref> system is definitely the most well known system. Another commercial system for image and video retrieval is Virage <ref type="bibr">[Bach 96] [Hampapur97]</ref> that has well known commercial customers such as CNN. In the academia, some systems including Candid <ref type="bibr" target="#b5">[Kelly95]</ref>, Photobook <ref type="bibr" target="#b6">[Pentland96]</ref>, and Netra <ref type="bibr" target="#b7">[Ma97]</ref> use simple color and texture characteristics to describe image content. The Blobword system <ref type="bibr" target="#b8">[Carson99]</ref>[Belongie98] exploits higher-level information, such as segmented objects of images, for queries. A system that is available free of charge is the GNU Image Finding Tool (GIFT) <ref type="bibr">[Squire00]</ref>. Some systems are available as demonstration version on the Web such as Viper, WIPE or Compass. Most of the available systems are hard to compare.</p><p>Imaging systems and image archives have often been described as an important economic and clinical factor in the hospital environment <ref type="bibr">[Greenes00]</ref>. Several methods from computer vision and image processing have already been proposed for the use in medicine <ref type="bibr" target="#b9">[Pun94]</ref>. Medical images have often been used for retrieval systems, and the medical domain is often cited as one of the principal application domains for content-based access technologies <ref type="bibr" target="#b0">[Smeul00]</ref> [Kelly95] <ref type="bibr" target="#b10">[Beretti01]</ref> [Orphan94] in terms of potential impact. Still, there has rarely been an evaluation of the performance.</p><p>One of the most significant problems in content-based image retrieval results from the lack of a common test-bed for researchers. Although many published articles report on content-based retrieval results using color photographs, there has been little effort in establishing a benchmark set of images and queries. It is very important that image databases are made available free of charge for the comparison and verification of algorithms. Only such reference databases allow comparing systems and to have a reference for the evaluation that is done based on the same images. ImageCLEF offers numerous medical images for evaluation that has many benefits in advancing the technology and utilization of content-based image retrieval systems.</p><p>In this year's ImageCLEF evaluation, we participated in the medical retrieval task. In the following sections, we detail the approach taken for the medical retrieval task. We analyze the results of the various evaluations, and have a discussion about the relative performance of our systems. In this task, we need to find the relevant images from approximate 9000 medical images that are similar to the 26 query images respectively. The data of this task contains cross language text (French and English) and visual medical images (CT, radiograph, MRI, and so on). We submitted two runs in this task. In the first run we use the visual features to retrieve similar images. In the second run we analyze the results of visual example queries and exploit a relevance feedback mechanism to improve the result. This paper is organized as follows. . In Section 2, we describe the features we use to represent the images. The similarity metric is proposed in Section 3. In Section 4, we explicate the automatic feedback mechanism. The submit runs are discussed in Section 5. Section 6 concludes this paper.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Feature Extraction</head><p>The medical image collection of ImageCLEF 2004 contains gray and color images. In color images, users are usually attracted by the change of colors more than the positions of objects. Thus, we use color histogram as the feature of color images to retrieve similar color images. Color histogram is suitable to compare images in many applications. Color histogram is computationally efficient, and generally insensitive to small changes in the camera position.</p><p>Color histogram has some drawbacks. Color histogram provides less spatial information; it merely describes which colors are present in an image, and in what quantities. Because gray images encompass few colors (usually 256 gray levels), directly using color histogram in gray images will get bad retrieval results. For gray images, we must emphasize spatial relationship analysis; furthermore, object and contrast analysis is important for medical images; therefore, three kinds of features that can indicate the spatial, coherence, and shape characteristics, gray-spatial histogram, coherence moment, and gray correlogram, are employed as the features of gray images.</p><p>In the following we describe the four kinds of features, one for color images and three for gray images, used in this paper.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1.">Color image features</head><p>Color histogram <ref type="bibr" target="#b1">[Swain91]</ref> is a basic method and has good performance for representing image content. The color histogram method gathers statistics about the proportion of each color as the signature of an image. Let C be a set of colors, (c 1 , c 2 …c m ) ∈ C, that can occur in an image. Let I be an image that consists of pixels p(x,y)<ref type="foot" target="#foot_0">1</ref> . The color histogram H(I) of image I is a vector (h 1 , h 2 , …, h i ,…, h m ), in which each bucket h i counts the ratio of pixels of color c i in I. Suppose that p is the color level of a pixel. Then the histogram of I for color c i is defined as Eq. (1):</p><formula xml:id="formula_0">} { Pr ) ( i I p c c p I h i ∈ = ∈<label>(1)</label></formula><p>In other words, corresponds to the probability of any pixel in I being of the color c ) (I h i c i . For comparing the similarity of two images I and I', the distance between the histograms of I and I' can be calculated using a standard method (such as the L 1 distance or L 2 distance). Then, the image in the image database most similar to a query image I is the one having the smallest histogram distance with I. Any two colors have a degree of similarity. Color histogram is hard to express the similar characteristic. In this paper, each pixel does not only assign a single color. We set an interval range δ to extend the color of each pixel. Then the histogram of image I is redefined as the Eq. ( <ref type="formula">2</ref>):</p><formula xml:id="formula_1">m c p p I h m j i j j c i ∑ = ∩ + − = 1 ] 2 , 2 [ ) ( δ δ δ (2)</formula><p>where p j is a pixel of image, and m is the total number of pixels.</p><p>The colors of an image are represented in HSV (Hue, Saturation, Value) space, which is closer to human perception than spaces such as RGB (Red, Green, Blue) or CMY (Cyan, Magenta, Yellow). In implementation, we quantize HSV space into 18 hues, 2 saturations and, 4 values, with additional 4 levels of gray values; as a result, there are a total of 148 bins.</p><p>Using the modified color histogram, the similarity of two color images q and d is defined as Eq. (3):</p><formula xml:id="formula_2">∑ ∑ = = = ∩ = n i i n i i i q h d h q h 1 1 ) ( )) ( ), ( min( | H(q) | H(d) H(q) H(d)) (q), SIMcolor(H<label>(3)</label></formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.">Gray image features</head><p>Gray images are different from color images in human's perception. Gray images have fewer colors than color images, only 256 gray levels in each gray image. Human's visual perception is influenced by the contrast of an image. The contrast of an image from the viewpoint of human is relative rather than absolute. To emphasize the contrast of an image and handle images with less illuminative influence, we normalize the value of pixels before quantization. In this paper we propose a relative normalization method. First, we cluster the whole image into four clusters by the K-means cluster method <ref type="bibr" target="#b11">[Han01]</ref>. We sort the four clusters ascendently according to their mean values. We shift the mean of the first cluster to value 50 and the fourth cluster to value 200; then each pixel in a cluster is multiplied by a relative weight to normalize. Let m c1 is the mean value of cluster 1 and m c4 is the mean value of cluster 4. The normalization formula of pixel p(x,y) is defined as Eq. (4).</p><formula xml:id="formula_3">) ( 200 )) 50 ( ) , ( ( ) , ( 1 4 1 c c c normal m m m y x p y x p − × − − =<label>(4)</label></formula><p>After normalization, we resize each image into 128*128 pixels, anduse one level wavelet with Haar wavelet function <ref type="bibr" target="#b12">[Stollnitz96]</ref> to generate the low frequency and high frequency sub-images. Process an image using the low pass filter will obtain an image that is more consistent than the original one; on the contrary, processing an image using the high pass filter will obtain an image that has high variation. The high-frequency part keeps the contour of the image. Figure <ref type="figure" target="#fig_0">1</ref> is an example of wavelet translation. Figure <ref type="figure" target="#fig_0">1</ref>   </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.1.">Gray-spatial histogram</head><p>In a gray image the spatial relationship is very important especially in medical images. Medical images always contain particular anatomic regions (lung, liver, head, and so on); therefore, similar images have similar spatial structures. We add spatial information into the histogram so we call this representation as gray-spatial histogram in order to distinguish from color histogram. We use the LL band for gray-spatial histogram and coherence analysis. To get the gray-spatial histogram, we divide the LL band image into nine areas. The gray values are quantized into 16 levels for computational efficiency.</p><p>The gray-spatial feature estimates the probability of each gray level that appears in a particular area. The probability equation is defined in Eq. (2), where δ is set to 10. The gray-spatial histogram of an image has a total of 144 bins.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.2.">Coherence moment</head><p>One of the problems to design a image representation is the semantic gap. The state-of-the-art technology still cannot reliably identify objects. The coherence moment feature attempts to describe the features from the human's viewpoint in order to reduce the semantic gap.</p><p>We cluster an image into four classes by the K-means algorithm. Figure <ref type="figure">2</ref> is an example. Figure <ref type="figure">2</ref> (a) is the original image and Figure <ref type="figure">2 (b</ref>) is four-level gray image. We almost can not visually find the difference between the two images. After clustering an image into four classes, we calculate the number of pixels (COH κ ), mean value of gray value (COH µ ) and standard variance of gray value (COH ρ ) in each class. For each class, we group connected pixels in eight directions as an object. If an object is bigger than 5% of the whole image, we denote it as a big object; otherwise it is a small object. We count how many big objects (COH ο ) and small objects (COH ν ) in each class, and use COH ο and COH ν as parts of image features.</p><p>Since we intend to know the reciprocal effects among classes, so we smooth the original image. If two images are similar, they will also be similar after smoothing. If their spatial distributions are quite different, they may have different result after smoothing. After smoothing, we cluster an image into four classes and calculate the number of big objects (COH τ ) and small objects (COH ω ). Figure <ref type="figure">3</ref> is an example. Each pixel will be influenced by its neighboring pixels. Two close objects of the same class may be merged into one object. Then, we can analyze the variation between the two images before and after smoothing. The coherence moment of each class is a seven-feature vector, (COH κ , COH µ , COH ρ , COH ο , COH ν , COH τ , COH ω ). The coherence moment of an image is a 28-feature vector that combines the coherence moments of the four classes.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.3.">Gray correlogram</head><p>The contour of a medical image contains rich information. Diseases can be easily detected in the high frequency domain. But, in this task we are going to find similar medical images, not to detect the affected part. A broken bone in the contour may be different from the healthy one. Thus we choose a representation that can estimate the partial similarity of two images and can be easy to calculate their global similarity.</p><p>We analyze the high frequency part by our modified correlogram algorithm. The definition of the correlogram [Huang97][Ojala01] is as Eq. ( <ref type="formula">5</ref>). Let D denote a set of fixed distances {d 1 , d 2 , d 3 ,…, d n }. The correlogram of an image I is defined as the probability of a color pair (ci, cj) at a distance d.</p><formula xml:id="formula_4">} | | { Pr ) ( 2 1 2 , , 2 1 d p p c p I j I p c p d c c i j i = − ∈ = ∈ ∈ γ (5)</formula><p>For computational efficiency, the autocorrelogram is defined as Eq. ( <ref type="formula">6</ref>)</p><formula xml:id="formula_5">} | | { Pr ) ( 2 1 2 , 2 1 d p p c p I i I p c p d c i i = − ∈ = ∈ ∈ λ (6)</formula><p>The contrast of a gray image dominates human's perception. If two images have different gray levels they still may be visually similar. Thus the coorelogram method cannot be used directly.</p><p>Our modified correlogram algorithm works as follows. First we sort the pixels of the high frequency part descendently.. Then we order the results of the preceding sorting by the ascendant distances of pixels to the center of the image. The distance of a pixel to the image center is measured by the L2 distance. After sorting by gray value and distance to the image center, we select the top 20 percent of pixels and the gray values higher than a threshold to estimate the autocorrelogram histogram. We set the threshold zero in this task. Any two pixels have a distance, and we estimate the probability that the distance falls within an interval. [76,100]}. The high frequent part comprises 64*64 pixels, thus the maximum distance will be smaller than 100. The first n pixels will have n*(n+1)/2 numbers of distances. We calculate the probability of each interval to form the correlogram vector.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Similarity metric</head><p>While an image has features to represent it, we need a metric to measure the similarity between two feature vectors (and consequently, the similarity between two images). The similarity metric of color histogram is defined as Eq. (3) and that of gray-spatial histogram is defined as Eq. ( <ref type="formula" target="#formula_6">7</ref>):</p><formula xml:id="formula_6">∑ ∑ = = = ∩ = n i i n i i i q h d h q h 1 1 al gray_spati ) ( )) ( ), ( min( | H(q) | H(d) H(q) H(d)) (H(q), SIM<label>(7)</label></formula><p>The similarity metric of the coherence moment is defined as Eq. ( <ref type="formula">8</ref></p><formula xml:id="formula_7">) |) ) ( ) ( | 2 | | | ) ( ) ( | | | | | | | | (| )) ( ), ( ( 2 / 1 2 / 1 2 / 1 2 / 1 4 1 i i i i i i i i i i i i i i d q d q d v q v d o q o classes i d q d q d q coh COH COH COH COH COH COH COH COH COH COH COH COH COH COH d COH q COH DIS ω ω τ τ ρ ρ µ µ κ κ − + × − + − + − + − × − + − = ∑ = (8)</formula><p>The correlogram metric is defined as Eq. ( <ref type="formula" target="#formula_8">9</ref>):</p><formula xml:id="formula_8">∑ ∑ = = + − = n i i i n i i i d h q h d h q h 1 1 hf | ) ( ) ( | | ) ( ) ( | H(d)) (H(q), DIS<label>(9)</label></formula><p>The similarity of two images Q and D is measured by Eq. (10):</p><formula xml:id="formula_9">SIM image (Q, D) = W 1 ×SIM color (H(Q),H(D) + W 2 ×SIM gray-spatial (H(Q),H(D) + W 3 ×1/(1+DIS coh (COH(Q),COH(D))) + W 4 × 1/(1+DIS hf (COH(Q),COH(D)))<label>(10)</label></formula><p>Figure <ref type="figure">4</ref> The feedback mechanism</p><p>, where W i is the weight of each feature. In this task the database contains color and gray images. When the user queries an image by example, we first determine whether the example is color or gray. We calculate the color histogram, if the four bins of gray values occupy more than 80% of the whole image, we decide the query image is gray; otherwise it is color. If the input is a color image, then we set W 1 =10, W 2 =0.1, W 1 =10, and W 1 =10; otherwise we set W 1 =0.1, W 2 =1, W 1 =100, and W 1 =100.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Feedback Mechanism</head><p>When the user inputs the visual query example, the system first employs visual features to retrieve relevant images from the database. After the initial retrieval, the system selects the top-n relevant images as candidate positive images. The similarity between the visual query example and each of the top-n images must also be greater than a threshold. In the next step, we cluster the top-n images into k classes. Figure <ref type="figure">4</ref> illustrates the feedback mechanism. In addition to images, the database of ImageCLEF 2004 contains a diagnosis text for each image. However, A patient case contains a variety of visual images. The images of the same case are sometimes not all similar in visually. So, in this paper while doing the relevance feedback, the weight of textual is slighter.</p><p>We first translate the diagnosis texts in French into possible English. The vector space model <ref type="bibr" target="#b13">[Salton88]</ref> is used to create a vector representation of a diagnosis text. Each entry of the vector represents a term of the text and the value of the entry is the term frequency (tf) * inverse document frequency (idf) value. The similarity between two diagnoses is computed as the cosine between their vector representations, as shown in Eq. ( <ref type="formula" target="#formula_9">10</ref>)</p><formula xml:id="formula_10">∑ ∑ ∑ = = = × × = n i d i n i q i n i d i q i w w w w d q ine 1 2 1 2 1 ) ( ) ( ) , ( cos<label>(10)</label></formula><p>where is the weight of term i in text q and n is the numbers of terms.</p><formula xml:id="formula_11">q i W</formula><p>The similarity between two images consists of visual similarity and textual similarity. We set the weight of textual part as 0.1 and visual features part as 0.9. In our implementation, we cluster the top-20 images into 6 classes by the minimum distance hierarchical cluster algorithm <ref type="bibr" target="#b11">[Han01]</ref>. The class most similar to the query example in vision becomes the next-generation query example images. We use the OR operation among exemplary images to measure the similarity of database images; in other words, we use the maximum similarity between positive query images and an image in the database to measure the similarity of the latter image to the query. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Experimental Results</head><p>In the ImageCLEF 2004, the process of evaluation and the format of results employs the trec_eval tool. The evaluation procedure did as following:</p><p>Extracted the top-60 runs from each submission (we received 43 submissions in total). Computed the union of runs to create a document pool for each topic. Manually assessed images in the document pool using three assessors (images judged as relevant and partially relevant). Created 9 sets of relevant images for each topic (9 qrels sets).</p><p>Compared each system run against the qrels. Computed uninterpolated mean average precision across all topics using trec_eval.</p><p>Three "expert" assessors judged the image pools generated from pooling the submissions. They created 9 sets of qrels based on the overlap of relevant images between assessors, and whether partially relevant images were included in the qrels set. The partially relevant judgment was used to pick up image where the judge thought it was in some way relevant, but could not be entirely confident. The 9 relevance sets are listed here: isec-rel: images judged as relevant by all three assessors. isec-partial: images judged as partially relevant by all three assessors. isec-total: images judged as either relevant or partially relevant by all three assessors. partial_isec-rel: images judged as relevant by at least 2 assessors. partial_isec-partial: images judged as partially relevant by at least 2 assessors. partial_isec-total: images judged as either relevant or partially relevant by at least 2 assessors. union_rel: images judged as relevant by all at least 1 assessor. union_partial: images judged as partially relevant by at least 1 assessor. union_total: images judged as either relevant or partially relevant by at least 1 assessor.</p><p>In this task, we submit three runs. We forgot to sort descendently the third run's output and get wrong results; thus we omit to discuss it. The first run uses the visual feature of the query example image to query the databse. The second run is the result of the automatic feedback mechanism, which uses the images of the most similar class as the positive query examples to query the image database. The test result shows that the auto-feedback mechanism, run2, has better result than the first run. In the partial_isec-total results summary, the mean average precision of the first run of our system is 0.2960. The mean average precision of run2 (with feedback) is 0.3457. Figure <ref type="figure" target="#fig_3">5</ref> shows the precision and recall graphs.  The first run has accuracy above 50% in the first 20 images. The really similar images may have similar features in some aspect and similar to each other. The miss judged images are always less consistent. So we try to refine the initial result by the automatic feed back mechanism. We cluster the first 20 images into six classes. If the class contains diverse images, the center of the class will become farther, and consequently more different, from the query image. Thus we can improve the result by our feedback method.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Conclusion</head><p>In this paper we propose several methods to represent medical images. Although the color histogram of content-based image retrieval methods has good performance in general-propose color images, unlike general-propose color images the X-ray images only contain gray-level pixels. Thus, we concentrate on the contrast representation of images.</p><p>The image representations we propose have obtained good results in this task. Our representation is immune in defective illumination. A total of 322 features is used. It is very efficient in computation. The auto feedback mechanism also provides a good result in medical images.</p><p>An image contains thousands of words. An image can be viewed from various aspects; furthermore, different people may have different interpretation of the same image. This will cause too many parameters need to be tuned. In the future, we will try to learn the user behavior and tune those parameters by artificial learning methods.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1</head><label>1</label><figDesc>Figure 1 An example of wavelet translation.</figDesc><graphic coords="3,246.96,72.78,119.34,96.00" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head></head><label></label><figDesc>(a) is the original image, (b) is one-level wavelet transformed image; there are four sub-bands denoted by Low_Low (LL), Low_Height (LH), Height_Low (HL), Height_Height (HH), as shown in Figure1 (c). High-frequency pixels may be important in medical images for doctor diagnoses. By performing the OR operation for LH, HL, and HH bands, we get the contour of a medical image.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head></head><label></label><figDesc>Figure 2 (a) original image with 256 levels; (b) new image after clustering with only 4 levels</figDesc><graphic coords="4,375.84,216.30,136.32,106.92" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Figure 5</head><label>5</label><figDesc>Figure 5 Precision Vs. Recall graphs without and with feedback</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_4"><head></head><label></label><figDesc>Figure 5 (a) is the run1 we submitted and Figure 5 (b) is the run2 with the automatic feedback mechanism.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_5"><head>Figure 6</head><label>6</label><figDesc>Figure 6 Result of an example query</figDesc><graphic coords="8,88.80,70.92,415.26,280.62" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0"><head></head><label></label><figDesc></figDesc><graphic coords="6,118.02,70.92,359.28,217.08" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head></head><label></label><figDesc>The distance intervals we set are {[0,2], [2,4], [4,6], [6,8], [8,12], [12,16], [16,26], [26,36], [36,46], [46,56], [56,66],</figDesc><table /></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">p(x, y) indicates the color of the corresponding pixel as well.</note>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Content-based image retrieval at the end of the early years</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">W M</forename><surname>Smeulders</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Worring</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Santini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Gupta</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Jain</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Pattern Analysis and Machine Intelligence</title>
		<imprint>
			<biblScope unit="volume">22</biblScope>
			<biblScope unit="issue">12</biblScope>
			<biblScope unit="page" from="1349" to="1380" />
			<date type="published" when="2000">2000</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Color Indexing</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">J</forename><surname>Swain</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">H</forename><surname>Ballard</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">International Journal of Computer Vision</title>
		<imprint>
			<biblScope unit="volume">7</biblScope>
			<biblScope unit="page" from="11" to="32" />
			<date type="published" when="1991">1991</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Query by Image and Video Content: The QBIC system</title>
		<author>
			<persName><forename type="first">M</forename><surname>Flickner</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Sawhney</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Niblack</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Ashley</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Huang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Dom</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Gorkani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Hafner</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Petkovic</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Steele</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Yanker</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Computer</title>
		<imprint>
			<biblScope unit="volume">28</biblScope>
			<biblScope unit="issue">9</biblScope>
			<biblScope unit="page" from="23" to="32" />
			<date type="published" when="1995">1995</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">The Virage image search engine: An open framework for image management</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">R</forename><surname>Bach</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Fuller</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Gupta</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Hampapur</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Horowitz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Humphrey</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Jain</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C.-F</forename><surname>Shu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Storage &amp; Retrieval for Image and Video Databases IV</title>
				<editor>
			<persName><forename type="first">I</forename><forename type="middle">K</forename><surname>Sethi</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">R</forename><forename type="middle">C</forename><surname>Jain</surname></persName>
		</editor>
		<meeting><address><addrLine>San Jose, CA, USA</addrLine></address></meeting>
		<imprint>
			<date type="published" when="1996">1996</date>
			<biblScope unit="volume">2670</biblScope>
			<biblScope unit="page" from="76" to="87" />
		</imprint>
	</monogr>
	<note>IS&amp;T/SPIE Proceedings</note>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Virage video engine</title>
		<author>
			<persName><forename type="first">A</forename><surname>Hampapur</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Gupta</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Horowitz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C.-F</forename><surname>Shu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Fuller</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Bach</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Gorkani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Jain</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Storage and Retrieval for Image and Video Databases V</title>
				<editor>
			<persName><forename type="first">I</forename><forename type="middle">K</forename><surname>Sethi</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">R</forename><forename type="middle">C</forename><surname>Jain</surname></persName>
		</editor>
		<imprint>
			<date type="published" when="1997">1997</date>
			<biblScope unit="volume">3022</biblScope>
			<biblScope unit="page" from="352" to="360" />
		</imprint>
	</monogr>
	<note>SPIE Proceedings</note>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Query by image example: the CANDID approach</title>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">M</forename></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Cannon</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">R</forename><surname>Hush</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Storage and Retrieval for Image and Video Databases III</title>
				<editor>
			<persName><forename type="first">W</forename><surname>Niblack</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">R</forename><forename type="middle">C</forename><surname>Jain</surname></persName>
		</editor>
		<imprint>
			<date type="published" when="1995">1995</date>
			<biblScope unit="volume">2420</biblScope>
			<biblScope unit="page" from="238" to="248" />
		</imprint>
	</monogr>
	<note>SPIE Proceedings</note>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Photobook: Tools for content-based manipulation of image databases</title>
		<author>
			<persName><forename type="first">A</forename><surname>Pentland</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">W</forename><surname>Picard</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Sclaro_</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">International Journal of Computer Vision</title>
		<imprint>
			<biblScope unit="volume">18</biblScope>
			<biblScope unit="issue">3</biblScope>
			<biblScope unit="page" from="233" to="254" />
			<date type="published" when="1996">1996</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Tools for texture-and color-based search of images</title>
		<author>
			<persName><forename type="first">W</forename><forename type="middle">Y</forename><surname>Ma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Deng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><forename type="middle">S</forename><surname>Manjunath</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Human Vision and Electronic Imaging II</title>
				<editor>
			<persName><forename type="first">B</forename><forename type="middle">E</forename><surname>Rogowitz</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">T</forename><forename type="middle">N</forename><surname>Pappas</surname></persName>
		</editor>
		<meeting><address><addrLine>San Jose, CA</addrLine></address></meeting>
		<imprint>
			<date type="published" when="1997">1997</date>
			<biblScope unit="volume">3016</biblScope>
			<biblScope unit="page" from="496" to="507" />
		</imprint>
	</monogr>
	<note>SPIE Proceedings</note>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Color and texture based image segmentation using EM and its application to content-based image retrieval</title>
		<author>
			<persName><forename type="first">C</forename><surname>Carson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Thomas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Belongie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">M</forename><surname>Hellerstein</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Malik ; Belongie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Carson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Greenspan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Malik</surname></persName>
		</author>
		<author>
			<persName><forename type="first">;</forename><surname>Muller</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Muller</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Pattern Recognition Letters (Selected Papers from The 11 th Scandinavian Conference on Image Analysis SCIA &apos;99)</title>
		<title level="s">Lecture Notes in Computer Science</title>
		<editor>
			<persName><forename type="first">R</forename><forename type="middle">A</forename><surname>Greenes</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><forename type="middle">F</forename><surname>Brinkley</surname></persName>
		</editor>
		<meeting><address><addrLine>Amsterdam, The Netherlands; Bombay, India; New York</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="1998">1999. 1998. 2000. 2000</date>
			<biblScope unit="volume">1614</biblScope>
			<biblScope unit="page" from="485" to="538" />
		</imprint>
	</monogr>
	<note>Medical Informatics: Computer Applications in Healthcare. 2nd edition</note>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Image analysis and computer vision in medicine</title>
		<author>
			<persName><forename type="first">T</forename><surname>Pun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Gerig</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Ratib</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Computerized Medical Imaging and Graphics</title>
		<imprint>
			<biblScope unit="volume">18</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page" from="85" to="96" />
			<date type="published" when="1994">1994</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">I2Cnet: A system for the indexing. storage and retrieval of medical images by content</title>
		<author>
			<persName><forename type="first">S</forename><surname>Beretti</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Del Bimbo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Pala</surname></persName>
		</author>
		<author>
			<persName><forename type="first">;</forename><forename type="middle">J</forename><surname>Huang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">R</forename><surname>Kumar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Mitra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W. -J</forename><surname>Zhu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Zabih</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Image indexing Using Color Correlograms, Proceedings of IEEE Conference on Computer Vision and Pattern Recognition</title>
				<editor>
			<persName><forename type="first">M</forename><surname>Matinmikko</surname></persName>
		</editor>
		<editor>
			<persName><surname>Aittola</surname></persName>
		</editor>
		<meeting><address><addrLine>Tokyo, Japan; San Juan, Puerto Rico; Bergen, Norway</addrLine></address></meeting>
		<imprint>
			<publisher>IEEE Computer Society</publisher>
			<date type="published" when="1994">2001. 1994. 1997. 2001</date>
			<biblScope unit="volume">19</biblScope>
			<biblScope unit="page" from="109" to="122" />
		</imprint>
	</monogr>
	<note>Semantic Image Retrieval with HSV Correlograms, Proceedings of 12 th Scandinavian Conference on Image Analysis</note>
</biblStruct>

<biblStruct xml:id="b11">
	<monogr>
		<title level="m" type="main">Data mining: concepts and techniques</title>
		<author>
			<persName><forename type="first">J</forename><surname>Han</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Kamber</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2001">2001</date>
			<publisher>Academic press</publisher>
			<pubPlace>San Diego, CA, USA</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<monogr>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">J</forename><surname>Stollnitz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">D</forename><surname>Derose</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">H</forename><surname>Salesin</surname></persName>
		</author>
		<title level="m">Wavelets for computer Graphics -Theory and Applications</title>
				<meeting><address><addrLine>San Francisco, CA</addrLine></address></meeting>
		<imprint>
			<publisher>Morgan Kaufmann Publilshers, Inc</publisher>
			<date type="published" when="1996">1996</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<monogr>
		<title level="m" type="main">Automatic Text Processing</title>
		<author>
			<persName><forename type="first">G</forename><surname>Salton</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1988">1988</date>
			<publisher>Addison-Wesley Publishing Company</publisher>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
