<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">MIRACLE&apos;s Combination of Visual and Textual Queries for Medical Images Retrieval</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Julio</forename><surname>Villena-Román</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">Universidad Carlos III de Madrid</orgName>
							</affiliation>
							<affiliation key="aff2">
								<orgName type="department">Decisions and Language</orgName>
								<orgName type="institution">DAEDALUS -Data</orgName>
								<address>
									<country>S.A</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">José</forename><forename type="middle">Carlos</forename><surname>González-Cristóbal</surname></persName>
							<affiliation key="aff1">
								<orgName type="institution">Universidad Politécnica de Madrid</orgName>
							</affiliation>
							<affiliation key="aff2">
								<orgName type="department">Decisions and Language</orgName>
								<orgName type="institution">DAEDALUS -Data</orgName>
								<address>
									<country>S.A</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">José</forename><forename type="middle">Miguel</forename><surname>Goñi-Menoyo</surname></persName>
							<affiliation key="aff1">
								<orgName type="institution">Universidad Politécnica de Madrid</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">José</forename><surname>Luís Martínez-Fernandez</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">Universidad Carlos III de Madrid</orgName>
							</affiliation>
							<affiliation key="aff2">
								<orgName type="department">Decisions and Language</orgName>
								<orgName type="institution">DAEDALUS -Data</orgName>
								<address>
									<country>S.A</country>
								</address>
							</affiliation>
						</author>
						<author role="corresp">
							<persName><forename type="first">Juan</forename><forename type="middle">José</forename><surname>Fernández</surname></persName>
							<email>jjfernandez@isys.dia.fi.upm.es</email>
							<affiliation key="aff1">
								<orgName type="institution">Universidad Politécnica de Madrid</orgName>
							</affiliation>
						</author>
						<title level="a" type="main">MIRACLE&apos;s Combination of Visual and Textual Queries for Medical Images Retrieval</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">D7747F52767D493F0E6E58E3855CB67F</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-25T00:35+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>H.3 [Information Storage and Retrieval]: H.3.1 Content Analysis and Indexing</term>
					<term>H.3.2 Information Storage</term>
					<term>H.3.3 Information Search and Retrieval</term>
					<term>H.3.4 Systems and Software. E.1 [Data Structures]. E.2 [Data Storage Representations]. H.2 [Database Management] Linguistic Engineering, Information Retrieval, image retrieval, cross-lingual, content-based retrieval, visual, text-based retrieval, textual, relevance feedback, GIFT, KSite, Lucene, IRMA project</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>This paper presents the 2005 MIRACLE's team participation in the ImageCLEFmed task of ImageCLEF 2005.</p><p>This task certainly requires the use of image retrieval techniques and therefore it is mainly aimed at image analysis research groups. Although our areas of expertise don't include image analysis research, we decided to make the effort to participate in this task to promote and encourage multidisciplinary participation in all aspects of information retrieval, no matter if it is text or content based. We resort to a publicly available image retrieval system (GIFT [1]) when needed.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>ImageCLEF is the cross-language image retrieval track which was established in 2003 as part of the Cross Language Evaluation Forum (CLEF), a benchmarking event for multilingual information retrieval held annually since 2000. Images are language independent by nature, but often they are accompanied by texts semantically related to the image (e.g. textual captions or metadata). Images can then be retrieved using primitive features based on its contents (e.g. visual exemplar) or abstract features expressed through text or a combination of both.</p><p>Originally, ImageCLEF focused specifically on evaluating the retrieval of images described by text captions using queries written in a different language, therefore having to deal with monolingual and bilingual image retrieval (multilingual retrieval was not possible as the document collection is only in one language) <ref type="bibr" target="#b14">[13]</ref>, <ref type="bibr" target="#b17">[14]</ref>. Later, the scope of ImageCLEF widened and goals evolved to investigate the effectiveness of combining text and image for retrieval (text and content-based) <ref type="bibr" target="#b6">[8]</ref>, collect and provide resources for benchmarking image retrieval systems and promote the exchange of ideas which will lead to improvements in the performance of retrieval systems in general.</p><p>With this objective in mind, a medical retrieval task was included in 2004 campaign <ref type="bibr" target="#b6">[8]</ref> and continued this year. In this task (referred as ImageCLEFmed), example images are used to perform a search against a medical image database consisting of images such as scans and x-rays <ref type="bibr" target="#b3">[4]</ref> to find similar images. Each medical image or a group of images represents an illness, and case notes in English or French are associated with each illness to be used for diagnosis or to perform a text-based query. Some of the queries are rather based on visual characteristics and responses with a content-based retrieval system may deliver satisfying results. Other queries cannot be solved with visual characteristics alone; thus they may seem very hard for visual-only retrieval researchers.</p><p>The MIRACLE team is made up of three university research groups located in Madrid (UPM, UC3M and UAM) along with DAEDALUS, a company founded in 1998 as a spin-off of two of these groups. DAEDALUS is a leading company in linguistic technologies in Spain and is the coordinator of the MIRACLE team. This is the third participation in CLEF, after years 2003 and 2004 <ref type="bibr" target="#b17">[14]</ref>, <ref type="bibr" target="#b11">[10]</ref>, <ref type="bibr" target="#b8">[9]</ref>, <ref type="bibr" target="#b2">[3]</ref>, <ref type="bibr" target="#b1">[2]</ref>. As well as bilingual, monolingual and cross lingual tasks, the team has participated in the ImageCLEF, Q&amp;A, WebCLEF and GeoCLEF tracks. This paper describes our participation in the ImageCLEFmed task of ImageCLEF 2005. This task certainly requires the use of image retrieval techniques and therefore it is mainly aimed at image analysis research groups. Although our areas of expertise don't include image analysis research, we decided to make the effort to participate in this task to promote and encourage multidisciplinary participation in all aspects of information retrieval, no matter if it is text or content based. We resort to a publicly available image retrieval system (GIFT <ref type="bibr" target="#b0">[1]</ref>) when needed.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Task goals</head><p>Image or multimedia retrieval is interesting for the domain of cross-language information retrieval as the media such as images are inherently almost insensitive to language. Many collections exist on the Internet which contain images as well as multilingual texts. However, the retrieval of images is an often-neglected topic in the information retrieval domain. In particular, hospitals produce an enormous amount of visual data but tools to manage these images and videos are scarce and exist currently only as research prototypes <ref type="bibr" target="#b3">[4]</ref>.</p><p>The main goal of ImageCLEFmed task is to improve the retrieval of medical images from heterogeneous and multilingual document collections containing images as well as text. The task is somewhat similar to the classic TREC ad hoc retrieval task, with a scenery in which a system knows the set of documents to be searched, but cannot anticipate the particular topic that will be investigated (i.e., topics are not known to the system in advance).</p><p>ImageCLEFmed 2005 extends the 2004 experiments with a larger database and more complex queries. The database consists of images <ref type="bibr" target="#b4">[5]</ref> from the Casimage (Radiology and pathology), MIR (Mallinckrodt Institute of Radiology, nuclear medicine), PEIR (Pathology Education Instructional Resource, Pathology and radiology) and PathoPIC (Pathology) datasets, with about 50,000 images in all. The collection also contains about 50,000 annotations in XML format. While the majority are written in English (over 40,000), there is a significant number in French (over 1,800) and German (over 7,800), and a few cases with no annotation at all. The quality of the texts is variable between collections and even within the same collection.</p><p>Query tasks have been formulated with example images and a short textual description explaining the research goal. The task organizers provide a list of topic statements in English, French and German, and a collection of images for each topic. Normally one or two example images for the desired result are supplied. One query also contains a negative example as a test. The goal of ImageCLEFmed is to retrieve as many relevant images as possible from the given visual and multilingual topics.</p><p>The task organizers have also made available results from a state-of-the-art image retrieval system (medGIFT) and a state-of-the-art text engine (Lucene [7]).</p><p>The next section is devoted to the description of the different experiments which were carried out.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Description of experiments</head><p>We focused our experiments to fully automatic retrieval, avoiding any manual feedback, and submitted runs both using only visual features for retrieval (content-based retrieval) and also runs using visual features and text (combination of content-based and text-based retrieval).</p><p>To isolate from the content-based retrieval part of the process, we resorted to GIFT (GNU Image Finding Tool) <ref type="bibr" target="#b0">[1]</ref>, a publicly available content-based image retrieval system which was developed under the GNU license and allows to perform query by example on images, using an image as the starting point for the search process. GIFT relies entirely on the image contents and thus it doesn't require the collection to be annotated. It also provides a mechanism to improve query results by relevance feedback.</p><p>Our approach is based on the multidisciplinary combination of GIFT content-based searches with text-based retrieval techniques. Our system consists of three parts: the content-based retrieval component (mainly GIFT), the text-based search engine and the merging component, which combines the results from the others to provide the final results.</p><p>We finally submitted 13 different runs to be evaluated by the task coordinators, which are explained in the following section.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Content-Based Retrieval</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Without feedback</head><p>This experiment consists on a content-based-only retrieval using GIFT. Initially the complete image database was indexed in a single collection using GIFT, down-scaling each image to 32x32 pixels. For each ImageCLEFmed query, a visual query is made up of all the images contained in the ImageCLEFmed query. Then, this visual query is introduced into the system to obtain the list of more relevant images (i.e., images which are more similar to those included in the visual query), along with the corresponding relevance values.. Although different search algorithms can be integrated as plug-ins in GIFT, only the provided separate normalisation algorithm has been used in our experiments.</p><p>There is only one submission, with "mirabase.qtop" as its run identifier.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>With feedback</head><p>These experiments are similar to the preceding one, also a content-based-only retrieval using GIFT, but incorporating relevance feedback. Each visual query is introduced into the system to obtain the list of images which are more similar to the visual query. Then the top N results are added to the original visual query to build a new visual query which is again introduced into the system to obtain the final list of results. In addition, GIFT allows to build a weighted visual query in which a relevance value may be associated to each included image.</p><p>There are 3 submissions:</p><formula xml:id="formula_0">• mirarf5.qtop</formula><p>This run takes the 5 most relevant images for feedback, each one with a value of 1 for its relevance in the visual query. The relevance in the visual query for the original images remains 1.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>• mirarf5.1.qtop</head><p>The same as mirarf5.qtop but using a value of 0.5 for the relevance in query of feedback images. The relevance in query for the original images remains 1.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>• mirarf5.2.qtop</head><p>The same as mirarf5.qtop but using a value of 0.5 for the relevance in query of the original images.</p><p>Finally, the different content-based runs are shown in Table <ref type="table" target="#tab_0">1</ref>. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Content-Based and Text-Based Mixed Retrieval</head><p>Our main interest is not in experiments where only image content is used in the retrieval process. Instead, our challenge was to test whether the text-based image retrieval could improve the analysis of the content of the image, or vice versa. We like to determine how text and image attributes can be combined to enhance image retrieval, in this case, in the medical domain.</p><p>First all the case annotations are indexed using a text-based retrieval engine (explained later). Natural language processing techniques are applied before indexing. An adhoc language-specific (for English, German and French) parser is used to identify different classes of alphanumerical tokens such as dates, proper nouns, acronyms, etc., as well as recognising common compound words. Text is tokenized, stemmed <ref type="bibr" target="#b12">[11]</ref>[12] and stop word filtered (for the three languages). Only one index is created, combining keywords in the three different languages.</p><p>Two different text-based retrieval engines were used. One was Lucene [7], with the results provided by the task organizers. The other engine was KSite <ref type="bibr" target="#b5">[6]</ref>, fully developed by DAEDALUS, which offers the possibility to use a probabilistic (BM25) model or a vector space model for the indexing strategy. Only the probabilistic model was used in our experiments.</p><p>The combination strategy consists on reordering the results from the content-based retrieval using a text-based retrieval. For each ImageCLEFmed query, a multilingual textual query is build with the English, German and French queries (first processing each one with the language depending parser and then concatenating the three lists), and executed in the search engine to obtain the list of top-1000 cases which are more relevant to the textual query.</p><p>The list of relevant images from the content-based retrieval is reordered, moving to the beginning of the list those images which belong to a case that is in the list of top-1000 cases. The rest of the images remain in the end of the list.</p><p>There are 10 different submissions:</p><p>• mirabasefil.qtop, mirarf5fil.qtop, mirarf5.1fil.qtop, mirarf5.2fil.qtop</p><p>These runs consist on the combination as previously described of content-based-only runs with the textbased retrieval obtained with KSite.</p><p>• mirabasefil2.qtop, mirarf5fil2.qtop, mirarf5.1fil2.qtop, mirarf5.2fil2.qtop</p><p>The same experiment, but using Lucene.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>• Other runs</head><p>Two other experiments were developed to test if there was any difference in results when using our own content-based GIFT index or using the medGIFT results provided by the task organizers. So, medGIFT was used as the starting point and then the same combination method as described before was applied. Finally, the different mixed retrieval runs are shown in Table <ref type="table" target="#tab_1">2</ref>. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Evaluation</head><p>Relevance assessments have been performed by experienced medical students and medical doctors at OHSU and the University hospitals of Geneva. Submissions from all groups are used to create image pools, which are judged for relevance by assessors, based on using a ternary classification scheme: (1) relevant, (2) partially relevant and (3) not relevant. The aim of the ternary scheme is to help assessors in making their relevance judgments more accurate (e.g., an image is definitely relevant in some way, but maybe the query object is not directly in the foreground; it is therefore considered partially relevant).</p><p>The pools are assessed and the end result is a set of relevance assessments called qrels, which are then used to evaluate system performance and compare submissions from different groups.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Content-Based runs</head><p>The results of the content-based runs are shown in the next table, ordered by its mean average precision value. As shown in Table <ref type="table" target="#tab_2">3</ref>, the best result was obtained with the base experiment, which means that relevance feedback has failed to improve the results (neither to worsen them). This may be due to an incorrect choice of the parameters, but this has to be further studied.</p><p>Apart from MIRACLE, other 8 groups participated in this year's evaluation in the content-based-only runs. Table <ref type="table" target="#tab_3">4</ref> compares each group's best submission. Only one group is above us in the group ranking, although their average precision is much better than ours. Our pragmatic approach using a "standard" publicly available content-based retrieval engine such as GIFT has proved to be a better approach than other presumably more complex techniques. We still have to test if another selection of indexing parameters (different from image down-scaling to 32x32 pixels and separate normalization algorithm) may provide better results.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Content-Based and Text-Based Mixed Retrieval</head><p>The results of the content-based and text-based mixed retrieval runs are shown in Table <ref type="table" target="#tab_4">5</ref>, ordered by its mean average precision value. In this case, using relevance feedback provides slightly better precision values.</p><p>Considering the best runs, the optimum choice seems to be to assume 1.0 for the relevance of the top 5 results and reduce the relevance of the images in the original query. Table <ref type="table" target="#tab_4">5</ref> also shows that the results are better with our own text-based search engine than using Lucene (all runs offer better precision values), at least with the adopted combination strategy. This difference could be attributed to a better language dependent pre-processing and removal of stop words.</p><p>It is interesting to observe that the worst combination is to take both results provided by the task organizers (content-based medGIFT results and text-based Lucene results), with a performance decrease of 15%.</p><p>Comparing content-based runs with the mixed runs, Table <ref type="table" target="#tab_5">6</ref> shows that the combination of both types of retrieval offers better performance and even the worst mixed run is better than the best content-based only run. This actually proves that text-based image retrieval can be used to improve the content-based only retrieval, with much superior performance. In this case, our position in the table shows that the submissions from other groups clearly surpassed our results. Anyway, these results are very satisfying for us, considering that we are not a group with expertise in image analysis research.</p><p>It is also interesting to note that most groups managed to improve their results with mixed approaches over the content-based only runs. This is especially visible for the NCTU group, with an improvement from 0.06 to 0.23 (+355%) in the average precision.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Conclusions and Future Work</head><p>Our main interest is not in experiments where only image content is used in the retrieval process. Instead, our challenge was to test whether the text-based image retrieval could improve the analysis of the content of the image, or vice versa. Results show that this hypothesis was right. Our combination of a "black-box" search using a publicly accessible content-based retrieval engine with a text-based search has turned to provide comparable results to other presumably "more complex" techniques. This simplicity may be a good starting point for the implementation of a real system.</p><p>We think that there still may be some space for improvement with a more careful study of the parameters for the relevance feedback and the combination strategy.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>o</head><label></label><figDesc>mirabase2fil.qtop medGIFT results filtered with text-based KSite results o mirabase2fil2.qtop medGIFT results filtered with Lucene results</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1 : Summary of content-based runs Run id Feedback Number of feedback images Relevance of original images Relevance of feedback images</head><label>1</label><figDesc></figDesc><table><row><cell>mirabase.qtop</cell><cell>NO</cell><cell>Not applicable</cell><cell>Not applicable</cell><cell>Not applicable</cell></row><row><cell>mirarf5.qtop</cell><cell>YES</cell><cell>5</cell><cell>1</cell><cell>1</cell></row><row><cell>mirarf5.1.qtop</cell><cell>YES</cell><cell>5</cell><cell>1</cell><cell>0.5</cell></row><row><cell>mirarf5.2.qtop</cell><cell>YES</cell><cell>5</cell><cell>0.5</cell><cell>1</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2 : Summary of mixed retrieval runs</head><label>2</label><figDesc></figDesc><table><row><cell>Run id</cell><cell>Feedback</cell><cell>Number of</cell><cell>Relevance of</cell><cell>Relevance of</cell><cell>Text</cell></row><row><cell></cell><cell></cell><cell>feedback images</cell><cell>original images</cell><cell>feedback images</cell><cell>retrieval engine</cell></row><row><cell>mirabasefil.qtop</cell><cell>NO</cell><cell>Not applicable</cell><cell>Not applicable</cell><cell>Not applicable</cell><cell>KSite</cell></row><row><cell>mirarf5fil.qtop</cell><cell>YES</cell><cell>5</cell><cell>1</cell><cell>1</cell><cell>KSite</cell></row><row><cell>mirarf5.1fil.qtop</cell><cell>YES</cell><cell>5</cell><cell>1</cell><cell>0.5</cell><cell>KSite</cell></row><row><cell>mirarf5.2fil.qtop</cell><cell>YES</cell><cell>5</cell><cell>0.5</cell><cell>1</cell><cell>KSite</cell></row><row><cell>mirarf5fil2.qtop</cell><cell>YES</cell><cell>5</cell><cell>1</cell><cell>1</cell><cell>Lucene</cell></row><row><cell>mirarf5.1fil2.qtop</cell><cell>YES</cell><cell>5</cell><cell>1</cell><cell>0.5</cell><cell>Lucene</cell></row><row><cell>mirarf5.2fil2.qtop</cell><cell>YES</cell><cell>5</cell><cell>0.5</cell><cell>1</cell><cell>Lucene</cell></row><row><cell>mirabase2fil.qtop</cell><cell>?</cell><cell>?</cell><cell>?</cell><cell>?</cell><cell>KSite</cell></row><row><cell>mirabase2fil2.qtop</cell><cell>?</cell><cell>?</cell><cell>?</cell><cell>?</cell><cell>Lucene</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 3 : Evaluation of content-based runs</head><label>3</label><figDesc></figDesc><table><row><cell>Run</cell><cell>Average</cell><cell>%</cell></row><row><cell></cell><cell>Precision</cell><cell></cell></row><row><cell>mirabase.qtop</cell><cell>0.0942</cell><cell></cell></row><row><cell>mirarf5.1.qtop</cell><cell cols="2">0.0942 100.0%</cell></row><row><cell>mirarf5.qtop</cell><cell>0.0941</cell><cell>99.8%</cell></row><row><cell>mirarf5.2.qtop</cell><cell>0.0934</cell><cell>99.1%</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 4 : Comparison of content-based runs from different groups</head><label>4</label><figDesc></figDesc><table><row><cell>Group</cell><cell>Average</cell><cell>%</cell></row><row><cell></cell><cell>Precision</cell><cell></cell></row><row><cell>I2R</cell><cell>0.1455</cell><cell></cell></row><row><cell>MIRACLE</cell><cell cols="2">0.0942 64.7%</cell></row><row><cell>GE</cell><cell cols="2">0.0941 64.7%</cell></row><row><cell>rwth</cell><cell cols="2">0.0751 51.6%</cell></row><row><cell>i6</cell><cell cols="2">0.0713 49.0%</cell></row><row><cell>nctu</cell><cell cols="2">0.0672 46.2%</cell></row><row><cell>&lt;unknown&gt;</cell><cell cols="2">0.0637 43.8%</cell></row><row><cell>ceamdI</cell><cell cols="2">0.0465 32.0%</cell></row><row><cell>cindi</cell><cell>0.0072</cell><cell>4.9%</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_4"><head>Table 5 : Comparison of mixed retrieval runs</head><label>5</label><figDesc></figDesc><table><row><cell>Run</cell><cell>Average Precision</cell><cell>%</cell></row><row><cell>mirarf5.2fil.qtop</cell><cell>0.1173</cell><cell></cell></row><row><cell>mirarf5fil.qtop</cell><cell cols="2">0.1171 99.8%</cell></row><row><cell>mirabasefil.qtop</cell><cell cols="2">0.1164 99.2%</cell></row><row><cell>mirabase2fil.qtop</cell><cell cols="2">0.1162 99.0%</cell></row><row><cell>mirarf5.1fil.qtop</cell><cell cols="2">0.1159 98.8%</cell></row><row><cell>mirarf5fil2.qtop</cell><cell cols="2">0.1028 87.6%</cell></row><row><cell>mirarf5.2fil2.qtop</cell><cell cols="2">0.1027 87.6%</cell></row><row><cell>mirarf5.1fil2.qtop</cell><cell cols="2">0.1019 86.9%</cell></row><row><cell>mirabasefil2.qtop</cell><cell cols="2">0.0998 85.1%</cell></row><row><cell>mirabase2fil2.qtop</cell><cell cols="2">0.0998 85.1%</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_5"><head>Table 6 : Comparison of content-based and mixed retrieval strategies</head><label>6</label><figDesc>Apart from MIRACLE, other 6 groups participated in this year's evaluation in the content-based and text-based runs. Table7shows the results for each group's best submission.</figDesc><table><row><cell>Run</cell><cell>Average</cell><cell>%</cell></row><row><cell></cell><cell>Precision</cell><cell></cell></row><row><cell>mirarf5.2fil.qtop</cell><cell>0.1173</cell><cell></cell></row><row><cell>mirabase2fil2.qtop</cell><cell cols="2">0.0998 85.1%</cell></row><row><cell>mirabase.qtop</cell><cell cols="2">0.0942 80%</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_6"><head>Table 7 : Comparison of mixed retrieval runs from other groups</head><label>7</label><figDesc></figDesc><table><row><cell>Group</cell><cell>Average</cell><cell>%</cell></row><row><cell></cell><cell>Precision</cell><cell></cell></row><row><cell>I2R</cell><cell>0.2821</cell><cell></cell></row><row><cell>Nctu</cell><cell cols="2">0.2389 84.7%</cell></row><row><cell>Ubimed</cell><cell cols="2">0.2358 83.6%</cell></row><row><cell>MIRACLE</cell><cell cols="2">0.1173 41.6%</cell></row><row><cell>GE</cell><cell cols="2">0.0981 34.8%</cell></row><row><cell>i6</cell><cell cols="2">0.0667 23.6%</cell></row><row><cell>ceamdl</cell><cell cols="2">0.0538 19.1%</cell></row></table></figure>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgements</head><p>This work has been partially supported by the Spanish R+D National Plan, by means of the project RIMMEL (Multilingual and Multimedia Information Retrieval, and its Evaluation), TIN2004-07588-C03-01 and also by the European Union with the funding of NEDINE project in the e-Content programme. Special mention to our colleagues of the MIRACLE team should be done (in alphabetical order): Ana María García-Serrano, Ana González-Ledesma, José José Mª Guirao-Miras, Sara Lana-Serrano, Paloma Martínez-Fernández, Ángel Martínez-González, Antonio Moreno-Sandoval and César de Pablo Sánchez.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<ptr target="http://www.gnu.org/software/gift/[Visited18/07/2005" />
		<title level="m">The GNU Image-Finding Tool</title>
				<imprint/>
	</monogr>
	<note>GIFT</note>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">MIRACLE&apos;s Hybrid Approach to Bilingual and Monolingual Information Retrieval</title>
		<author>
			<persName><surname>Goñi-Menoyo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M;</forename><surname>José</surname></persName>
		</author>
		<author>
			<persName><surname>González</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>José</surname></persName>
		</author>
		<author>
			<persName><forename type="first">José</forename><forename type="middle">L</forename><surname>Martínez-Fernández</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Villena</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">CLEF 2004 proceedings</title>
		<title level="s">Lecture Notes in Computer Science</title>
		<editor>
			<persName><forename type="first">C</forename><surname>Peters</surname></persName>
		</editor>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2005">2005</date>
			<biblScope unit="volume">3491</biblScope>
			<biblScope unit="page" from="188" to="199" />
		</imprint>
	</monogr>
	<note>to appear</note>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">MIRACLE&apos;s hybrid approach to bilingual and monolingual Information Retrieval</title>
		<author>
			<persName><surname>Goñi-Menoyo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>José</surname></persName>
		</author>
		<author>
			<persName><surname>González</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>José</surname></persName>
		</author>
		<author>
			<persName><surname>Martínez-Fernández</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>José</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Julio</forename><forename type="middle">;</forename><surname>Villena-Román</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ana</forename><forename type="middle">;</forename><surname>García-Serrano</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Paloma</forename><forename type="middle">;</forename><surname>Martínez-Fernández</surname></persName>
		</author>
		<author>
			<persName><forename type="first">César</forename><forename type="middle">;</forename><surname>De Pablo-Sánchez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Alonso-Sánchez</forename></persName>
		</author>
		<author>
			<persName><forename type="first">Javier</forename></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Working Notes for the CLEF 2004 Workshop</title>
				<editor>
			<persName><forename type="first">Carol</forename><surname>Peters</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Francesca</forename><surname>Borri</surname></persName>
		</editor>
		<meeting><address><addrLine>Bath, United Kingdom</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2004">2004</date>
			<biblScope unit="page" from="141" to="150" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Image Information Retrieval: An Overview of Current Research</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">A</forename><surname>Goodrum</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Informing Science</title>
		<imprint>
			<biblScope unit="volume">3</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page" from="63" to="66" />
			<date type="published" when="2000">2000</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<monogr>
		<ptr target="http://www.irma-project.org/[Visited18/07/2005" />
		<title level="m">Image Retrieval in Medical Applications</title>
				<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<monogr>
		<title level="m" type="main">Agente Corporativo</title>
		<author>
			<persName><surname>Ksite</surname></persName>
		</author>
		<ptr target="http://www.daedalus.es/ProdKSiteAC-E.php" />
		<imprint>
			<date type="published" when="2005-07-13">13/07/2005</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<monogr>
		<title/>
		<author>
			<persName><surname>Martínez-Fernández</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>José</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ana</forename><surname>García-Serrano</surname></persName>
		</author>
		<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">MIRACLE approach to ImageCLEF 2004: merging textual and content-based Image Retrieval</title>
		<author>
			<persName><forename type="first">J</forename><surname>Villena</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Méndez-Sáez</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">CLEF</title>
		<title level="s">Lecture Notes in Computer Science</title>
		<editor>
			<persName><forename type="first">C</forename><surname>Peters</surname></persName>
		</editor>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2004">2004. 2005</date>
			<biblScope unit="volume">3491</biblScope>
		</imprint>
	</monogr>
	<note>to appear</note>
</biblStruct>

<biblStruct xml:id="b8">
	<monogr>
		<title/>
		<author>
			<persName><forename type="first">José</forename><forename type="middle">L</forename><surname>Martínez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Julio</forename><surname>Villena</surname></persName>
		</author>
		<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<monogr>
		<title/>
		<author>
			<persName><forename type="first">Jorge</forename><surname>Fombella</surname></persName>
		</author>
		<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">MIRACLE Approaches to Multilingual Information Retrieval: A Baseline for Future Research. Comparative Evaluation of Multilingual Information Access Systems</title>
		<author>
			<persName><forename type="first">G</forename><surname>Serrano</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ana</forename><forename type="middle">;</forename><surname>Martínez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Paloma</forename><forename type="middle">;</forename><surname>Goñi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">José</forename><forename type="middle">M</forename><surname>González</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>José</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="s">Lecture Notes in Computer Science</title>
		<editor>Peters, C</editor>
		<editor>Gonzalo, J.</editor>
		<editor>Brascher, M.</editor>
		<editor>and Kluck, M.</editor>
		<imprint>
			<biblScope unit="volume">3237</biblScope>
			<biblScope unit="page" from="210" to="219" />
			<date type="published" when="2004">2004</date>
			<publisher>Springer</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Evaluation of MIRACLE approach results for CLEF</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">L</forename><surname>Martínez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Villena-Román</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Fombella</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>García-Serrano</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Ruiz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Martínez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">M</forename><surname>Goñi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">C</forename><surname>González</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ed</forename><surname>Carol Peters</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Working Notes for the CLEF 2003 Workshop</title>
				<meeting><address><addrLine>Trondheim, Norway</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2003-08">2003. August</date>
			<biblScope unit="page" from="21" to="22" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<monogr>
		<title level="m" type="main">Snowball stemmers and resources page</title>
		<author>
			<persName><forename type="first">Martin</forename><surname>Porter</surname></persName>
		</author>
		<ptr target="http://www.snowball.tartarus.org" />
		<imprint>
			<date type="published" when="2005-07-13">13/07/2005</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<monogr>
		<ptr target="Visited13/07/2005" />
		<title level="m">Page of resources for CLEF (Stopwords, transliteration, stemmers …</title>
				<imprint/>
		<respStmt>
			<orgName>University of Neuchatel</orgName>
		</respStmt>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<monogr>
		<title/>
		<author>
			<persName><forename type="first">Julio</forename><surname>Villena</surname></persName>
		</author>
		<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<monogr>
		<title/>
		<author>
			<persName><forename type="first">José</forename><forename type="middle">L</forename><surname>Martínez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jorge</forename><surname>Fombella</surname></persName>
		</author>
		<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Image Retrieval: The MIRACLE Approach. Comparative Evaluation of Multilingual Information Access Systems</title>
		<author>
			<persName><forename type="first">G</forename><surname>Serrano</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ana</forename><forename type="middle">;</forename><surname>Ruiz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Alberto</forename><forename type="middle">;</forename><surname>Martínez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Paloma</forename><forename type="middle">;</forename><surname>Goñi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">José</forename><forename type="middle">M</forename><surname>González</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>José</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="s">Lecture Notes in Computer Science</title>
		<editor>Peters, C</editor>
		<editor>Gonzalo, J.</editor>
		<editor>Brascher, M.</editor>
		<editor>and Kluck, M.</editor>
		<imprint>
			<biblScope unit="volume">3237</biblScope>
			<biblScope unit="page" from="621" to="630" />
			<date type="published" when="2004">2004</date>
			<publisher>Springer</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">MIRACLE results for ImageCLEF</title>
		<author>
			<persName><forename type="first">J</forename><surname>Villena-Román</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">L</forename><surname>Martínez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Fombella</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>García-Serrano</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Ruiz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Martínez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">M</forename><surname>Goñi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">C</forename><surname>González</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ed</forename><surname>Carol Peters</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Working Notes for the CLEF 2003 Workshop</title>
				<meeting><address><addrLine>Trondheim, Norway</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2003-08">2003. August</date>
			<biblScope unit="page" from="21" to="22" />
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
