<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Merging results from different media: Lic2m experiments at ImageCLEF 2005</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author role="corresp">
							<persName><forename type="first">Romaric</forename><surname>Besançon</surname></persName>
							<email>besanconr@zoe.cea.fr</email>
							<affiliation key="aff0">
								<orgName type="institution">Christophe Millet CEA-LIST</orgName>
								<address>
									<addrLine>LIC2M ; BP 6</addrLine>
									<postCode>92265</postCode>
									<settlement>Fontenay-aux-Roses CEDEX</settlement>
									<country key="FR">FRANCE</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Merging results from different media: Lic2m experiments at ImageCLEF 2005</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">D6FCC1E280572FEE1F5DB7BEE2420996</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-25T00:39+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>H.3 [Information Storage and Retrieval]: H.3.1 Content Analysis and Indexing</term>
					<term>H.3.3 Information Search and Retrieval</term>
					<term>H.3.4 Systems and Software</term>
					<term>H.3.7 Digital Libraries</term>
					<term>I.4.7 [Computing Methodologies]: Image Processing and Computer Vision-Feature Measurement Measurement, Performance, Experimentation Linguistic Processing, Cross-lingual Text Retrieval, Content Based Image Retrieval</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>In the ImageCLEF 2005 campaign, the LIC2M participated in the ad hoc task, the medical task and the annotation task. For both ad hoc and medical task, we perform experiments on merging the results of two independent search systems: a crosslanguage information retrieval system exploiting the text part of the query and a content-based image retrieval system exploiting the example images given with the query. The results show that a well-tuned merging may improve performance, but the tuning is made difficult because the performance of each system highly depends on the corpus and queries. Annotation task has been performed using a KNN classifier with the image indexes of our CBIR system.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>ImageCLEF campaign aims at studying cross-language image retrieval, that potentially uses text and image matching techniques. The LIC2M participated in ImageCLEF 2005 to perform experiments on merging strategies to integrate the results obtained from the cross-language text retrieval system and the content-based image retrieval system that are developed in our lab.</p><p>In both ad hoc and medical tasks of the ImageCLEF 2005 campaign, text and visual information were provided for the queries. In ad hoc task, the basic query is textual (title and narrative are provided), but two example images are provided; in medical task, query images are given and a short textual description give precisions about the research goal. We applied the same strategy for the two tasks, using our general-domain systems for multilingual text retrieval and contentbased image retrieval, taking into account both text and visual part of the query and applying a posteriori merging strategies on the results provided independently by each system. We present in section 2 the retrieval systems for text and image and the merging strategies used. We then present the results obtained for the ad hoc task and the medical task in sections 3 and 4 respectively. The strategy and results for the annotation task are presented in section 5.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Retrieval systems</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1">Multilingual Text Retrieval System</head><p>The multilingual text retrieval system used for these experiments is basically the same as the one used for the previous CLEF campaigns, and a more detailed description can be found in <ref type="bibr" target="#b0">[1]</ref>. The system has not been specially adapted to work on the text of the ImageCLEF corpora, and has simply been used as is. In particular, for both the ad hoc and medical corpora, no special treatment has been performed to take into account the structure of the documents (such as photographer's name, location, date for the captions and description, diagnosis, clinical presentation in the medical annotations): all fields containing some text have been taken as is. No adaptation has been made to take into account the specificities of medical texts (specialized vocabulary). Notice that this system is not only cross-lingual but multilingual, because it integrates a concept-based merging technique to merge results found in each target language. Its basic principle is briefly described here.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Document and query processing</head><p>The documents and queries are processed through a linguistic analyzer, that performs in particular a part-of-speech tagging, a lemmatization, and extracts compounds and named entities from the text. The elements extracted from the documents are indexed into inverted files. The elements extracted from the queries are used as query "concepts". Each concept is reformulated into a set of search terms for each target language, either using a monolingual expansion dictionary (that introduces synonyms and related words), or using a bilingual dictionary. Document Retrieval Each search term is searched in the index, and documents containing the term are retrieved. All retrieved documents are then associated with a concept profile, indicating the presence of query concepts in the document. This concept profile depends on the query concepts, and is language-independent (which allow merging results from different languages). Documents sharing the same concept profile are clustered together, and a weight is associated with each cluster according to its concept profile and to the weight of the concepts (the weight of a concept depends on the weight of each of its reformulated term in the retrieved documents). The clusters are sorted according to their weights and the first 1000 documents in this sorted list are retrieved.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2">Content-based Image Retrieval System</head><p>The content-based image retrieval system we used in ImageCLEF 2005 is the system PIRIA (Program for the Indexing and Research of Images by Affinity) <ref type="bibr" target="#b2">[3]</ref>, developed in our lab. The query image is submitted to the system, which returns a list of images ranked by their similarity to the query image. The similarity is obtained by a metric distance that operates on every image signatures. These indexed images are compared according to several classifiers : principally Color, Texture and Form if the segmentation of the images is relevant. The system takes into account geometric transformations and variations like rotation, symmetry, mirroring, etc. PIRIA is a global one-pass system, feedback or "relevant/non relevant" learning methods are not used.</p><p>Color Indexing This indexer first quantifies the image, and then, for each quantified color, it computes how much this color is connex. It can also be described as a border/interior pixel classification <ref type="bibr" target="#b3">[4]</ref>. The distance used for the color indexing is a classical L2 norm.</p><p>Texture Indexing A global texture histogram is used for the texture analysis. The histogram is computed from the Local Edge Pattern descriptors <ref type="bibr" target="#b1">[2]</ref>. These descriptors describe the local structure according to the edge image computed with a Sobel filtering. We obtain a 512-bins texture histogram, which is associated with a 64-bins color histogram where each plane of the RGB color space is quantized into 4 colors. Distances are computed with a L1 norm.</p><p>Form Indexing The form indexer used consists of a projection of the edge image along its horizontal and vertical axes. The image is first resized in 100x100. Then, the Sobel edge image is computed and divided into four equal sized squares (up left, up right, bottom left and bottom right). Then, each 50x50 part is projected along its vertical and horizontal axes, thus giving a 400-bins histogram. The L2 distance is used to compare two histograms.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.3">Search and Merging Strategy</head><p>For both ad hoc and medical task, the queries contain textual and visual information. Textual information is used to search relevant text documents with multilingual text retrieval system. For ad hoc task, each text document corresponds to a single image: the images corresponding to the relevant texts are then given as results. For the medical task, a text document may be associated with several images. In that case, the score obtained by the text documents is given to each image it is associated with: the first 1000 images in this image list are kept.</p><p>Since the PathoPic corpus of the medical task contains annotations in English and German that are associated with the same image, the multilingual retrieval system may return both English and German annotations as relevant documents (maybe with different scores), creating duplicate elements in the result list. In this case, the score associated with the corresponding image is the best score returned. To make sure that the number of retrieved images is 1000, we set the number of retrieved documents for the text retrieval system at 2000 for the medical task <ref type="foot" target="#foot_0">1</ref> .</p><p>Independently, visual information was used by the CBIR system to retrieve similar images. Queries contain several images: a first merging has been performed to obtain a single image list from the results of each query image: the score associated to result images is set to the max of the scores obtained for each query image.</p><p>Merging the results obtained by each system is simply done by a weighted sum of the scores obtained by each system. To be comparable, the scores of each system are normalized, for each query, by the highest score obtained for the query. This merging is parameterized by a merging coefficient α: for a query q and an image document retrieved for this query d ∈ Ret(q), the merging score is</p><formula xml:id="formula_0">s(d) = α× s T (d) max d∈RetT (q) s T (d) + (1 − α)× s I (d) max d∈RetI (q) s I (d)</formula><p>where s T (d) is the score of the text retrieval system and s I (d) the score of the image retrieval system.</p><p>A conservative merging strategy has also been tested: by conservative, we mean that we use the results obtained by one system only to reorder the results obtained by the other (results can be added at end of list if the number of documents retrieved by main system is less than 1000). The score of a document is modified using the same merging coefficient. For example, if the merging is conservative with the text results:</p><formula xml:id="formula_1">s (d) = s(d) if s T (d) = 0 0 otherwise</formula><p>The results we obtained in ImageCLEF 2004 tend to show that this kind of conservative merging strategies gives good performances. We will use the term of expansionist merging strategy to denote standard merging strategy, as opposed to the conservative one.</p><p>In the ad hoc task, we used textual queries in English, French and Spanish. We tried using the title only (T) or the title and the narrative (T+N). Comparative results for textual retrieval only, using either T or T+N are given in Table <ref type="table" target="#tab_0">1</ref>. These results show that average precision is better when using the title only, but the number of relevant documents is generally better when using also the narrative part (except for French, for which it is a bit worse). This can be explained by the fact that narrative introduce more words that allow to increase the total number of documents retrieved (for English and Spanish, there are 6 queries for which the system does not find 1000 documents matching title only, only 3 for French), and the number of relevant documents. But narrative also introduces more noise, which makes the precision decrease. We present in Table <ref type="table" target="#tab_1">2</ref> the results obtained by the merging of the two systems, using the texture indexer for the CBIR system. The results are presented for conservative and expansionist strategies and for different values of the merging coefficient α (when α = 1, the search is only based on text, when α = 0, the search is only based on images). Values below 0.5 are not presented but does not give better results. For expansionist strategy, the results are given for the mean average precision (map) and the number of relevant documents retrieved (relret); for conservative strategy, only the map is presented (relret is constant).</p><p>These results show that this simple merging of text and image results based on a weighted sum of the scores can increase the mean average precision (gain of 17 or 18%) and the best value for α is around 0.7 (though differences with surrounding values are small).</p><p>Concerning conservative/expansionist strategies, our previous experiments in ImageCLEF showed that the StAndrews collection, composed of old photographs, is not well adapted the kind of image indexers we use, that rely mostly on color for segmentation. We therefore chose the text retrieval as base for conservative merging. Looking at the relevant documents retrieved prove us right: text retrieval allow to retrieve 1246 relevant documents, whereas image retrieval only retrieve 367 relevant documents (239 of which were also found by the text retrieval system). However, the two merging strategies give comparable results, even though, as one can expect, the performance of the expansionist strategy decreases faster with α. Submitted runs for the ad hoc task in the ImageCLEF 2005 campaign were, for English, French and Spanish, text only (T and T+N), plus a conservative merging of text results based on title only and image results based on the texture indexer, with α = 0.9.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Results for the Medical task</head><p>In the medical task, we tested text retrieval using queries in English, French and German (searching for each in all target languages).</p><p>Based on our experiments in ImageCLEF 2004, we assumed that image retrieval for the medical task gives good results. Submitted runs for the medical task in the ImageCLEF 2005 campaign include runs based on visual queries only (texture and color indexers), and for English, French and German, a conservative merging of image results based on the texture indexer and text results, with α = 0.9. Unfortunately, the use of texture or color indexer with the ImageCLEFmed 2005 visual queries gave poor results, and conservative merging based on these results does not give much better results<ref type="foot" target="#foot_1">2</ref> .</p><p>We present in Table <ref type="table">4</ref> the results obtained by the merging of text and image systems, using the texture indexer for the CBIR system, with different values of the merging coefficient α, and for conservative and expansionist merging strategies (conservative strategy based on text results).</p><p>Except for German (for which our linguistic processing is clearly not well adapted to medical text), the conservative merging strategy improves performances (the best merging coefficient seems to be around 0.5). Expansionist merging gives comparable results: improvement of mean average precision is less important, but the number of relevant documents retrieved is generally improved, which tends to prove that both systems retrieve different documents<ref type="foot" target="#foot_2">3</ref> : conservative merging improves the ordering of documents retrieved by one system whereas expansionist merging improves the number of documents retrieved.</p><p>We present in Table <ref type="table" target="#tab_3">5</ref> similar results using the color indexer for visual retrieval. Results are slightly worse, but the same kind of tendencies as for the texture indexer can be noticed.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Annotation task</head><p>For the automatic annotation task, we submitted three runs, each corresponding to one of the three indexers described in section 2 (Color, Texture and Form).</p><p>All images are first indexed with the chosen indexer. Then, a k-Nearest Neighbor classifier is used to classify the indexed images. Odd numbers from <ref type="bibr" target="#b2">3</ref>  indexer, and evaluated with the leave-one-out method. The best k were 3 for the form indexer and 9 for the color indexer and the texture indexer.</p><p>The attributed class is decided by a majority vote of the nearest neighbors. In case of ties, distances to nearest neighbors are used (for example, in 9-NN, if 4 neighbors are from a class A, 4 neighbors from a class B, and 1 from another class, we use the distances between the requested image and its neighbors to select the nearest class).</p><p>We present in Table <ref type="table" target="#tab_4">6</ref> the results obtained for each of the indexers. It is not a surprise that the form indexer performed better than the others, as all the images in the database were in grey levels, and the form indexer is designed for such images, whereas the color and texture indexers are not well adapted to it (remember that the texture indexer includes a 64-bins color histogram). 9-NN Color 9-NN Texture 3-NN Form error rate 46.0 % 42.5 % 36.9 % </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6">Conclusion</head><p>The experiments performed by the LIC2M in the ImageCLEF 2005 campaign show that merging results from different media may increase the performance of a search system: a well-tuned a posteriori merging of the results obtained by two general purpose systems (no particular adaptation of the systems was made for the two tasks) can improve the mean average precision by at least 15%.</p><p>The difficulty relies on the tuning of the merging strategy. We used a simple weighted sum of the scores given by each system but the importance given to each system should rely on the performance of the system on a particular corpus, that is not easily predicted (best strategy for the ImageCLEF 2004 medical task appears to be opposite to the best strategy for ImageCLEF 2005 medical task, that has a more varied corpus and more difficult visual queries).</p><p>Further experiments will be undertaken to try to make the systems give a confidence score associated with its results and adapt the merging strategy according to this confidence. Other more sophisticated merging strategies will also be considered.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1 :</head><label>1</label><figDesc>Ad hoc task: comparative results using title or title+narrative: mean average precision (map), number of relevant documents (relret) and recall at 1000 documents (r1000 )</figDesc><table><row><cell></cell><cell></cell><cell>eng</cell><cell>fre</cell><cell></cell><cell>spa</cell><cell></cell></row><row><cell></cell><cell>T</cell><cell>T+N</cell><cell>T</cell><cell>T+N</cell><cell>T</cell><cell>T+N</cell></row><row><cell cols="2">map 0.246</cell><cell>0.224</cell><cell>0.186</cell><cell cols="2">0.146 0.191</cell><cell>0.151</cell></row><row><cell>relret</cell><cell>1246</cell><cell>1401</cell><cell>1237</cell><cell>1184</cell><cell>1085</cell><cell>1153</cell></row><row><cell>r1000</cell><cell>65%</cell><cell cols="5">73.1% 64.6% 61.8% 56.6% 60.2%</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2 :</head><label>2</label><figDesc>Ad hoc task: comparative results for the merging strategies using texture indexer for the image retrieval Similar results are presented in Table3using the color indexer for the CBIR system. Results are comparable: for this corpus, the two image indexers tend to retrieve similar documents (2/3 of relevant documents retrieved by both systems are identical).</figDesc><table><row><cell></cell><cell cols="3">conservative merging (map)</cell><cell></cell><cell cols="3">expansionist merging (map / relret)</cell></row><row><cell>α</cell><cell>eng</cell><cell>fre</cell><cell>spa</cell><cell>α</cell><cell>eng</cell><cell>fre</cell><cell>spa</cell></row><row><cell>1</cell><cell>0.246</cell><cell>0.186</cell><cell>0.191</cell><cell>1</cell><cell cols="3">0.246/ 1246 0.186 / 1237 0.191 / 1085</cell></row><row><cell cols="2">0.9 0.274</cell><cell>0.214</cell><cell>0.209</cell><cell cols="4">0.9 0.274/1254 0.214 / 1254 0.208 / 1088</cell></row><row><cell cols="4">0.8 0.282 0.221 0.212</cell><cell cols="4">0.8 0.28 / 1221 0.221/ 1251 0.212 / 1091</cell></row><row><cell cols="2">0.7 0.28</cell><cell>0.22</cell><cell>0.223</cell><cell cols="4">0.7 0.275/ 1214 0.204 /1254 0.222/ 1094</cell></row><row><cell cols="3">0.6 0.282 0.218</cell><cell>0.225</cell><cell cols="4">0.6 0.26 / 1166 0.176 / 1231 0.219 / 1086</cell></row><row><cell cols="2">0.5 0.276</cell><cell>0.213</cell><cell>0.227</cell><cell cols="4">0.5 0.236/ 1143 0.162 / 1146 0.217 /1098</cell></row><row><cell>0</cell><cell>0.068</cell><cell>0.068</cell><cell>0.068</cell><cell>0</cell><cell cols="3">0.068/ 367 0.068 / 367 0.068 / 367</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 3 :</head><label>3</label><figDesc>Ad hoc task: comparative results for the merging strategies using color indexer for the image retrieval</figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 5 :</head><label>5</label><figDesc>Medical task: comparative results for the merging strategies using color indexer for the image retrieval</figDesc><table><row><cell>to 13 have been tested for k for each</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_4"><head>Table 6 :</head><label>6</label><figDesc>Results for the automatic annotation task</figDesc><table /></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">this duplication of results was not detected before the submission of the runs, but the technique we used for merging text and image results remove the duplicate documents.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_1">Furthermore, we detected a bug in submitted runs, concerning the document identifier matching (1 vs. 0000001 ) that made the Peir corpus documents ignored in text retrieval results.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_2">We verified that text results with English queries contain 999 relevant images, image results with texture indexer contain 822 relevant images and only 218 images were common to the two systems.</note>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Concept-based searching and merging for multilingual information retrieval: First experiments at clef</title>
		<author>
			<persName><forename type="first">Romaric</forename><surname>Besançon</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Olivier</forename><surname>Gaël De Chalendar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Christian</forename><surname>Ferret</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Olivier</forename><surname>Fluhr</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Hubert</forename><surname>Mesnard</surname></persName>
		</author>
		<author>
			<persName><surname>Naets</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Comparative Evaluation of Multilingual Information Access Systems</title>
				<editor>
			<persName><forename type="first">Carol</forename><surname>Peters</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Julio</forename><surname>Gonzalo</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Martin</forename><surname>Braschler</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Michael</forename><surname>Kluck</surname></persName>
		</editor>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2003">2003. 2004</date>
			<biblScope unit="page" from="174" to="184" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Image classification using color, texture and regions</title>
		<author>
			<persName><forename type="first">Ya-</forename></persName>
		</author>
		<author>
			<persName><forename type="first">Chun</forename><surname>Cheng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Shu-Yuan</forename><surname>Chen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Image and Vision Computing</title>
		<imprint>
			<biblScope unit="volume">21</biblScope>
			<biblScope unit="issue">9</biblScope>
			<date type="published" when="2003-09">September 2003</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">PIRIA : A general tool for indexing, search and retrieval of multimedia content</title>
		<author>
			<persName><forename type="first">Magali</forename><surname>Joint</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Pierre-Alain</forename><surname>Moëllic</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Patrick</forename><surname>Hède</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Pascal</forename><surname>Adam</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">SPIE Electroning Imaging 2004</title>
				<meeting><address><addrLine>San Jose, California USA</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2004-01">January 2004</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">A compact and efficient image retrieval approach based on border/interior pixel classification</title>
		<author>
			<persName><forename type="first">Renato</forename><forename type="middle">O</forename><surname>Stehling</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Mario</forename><forename type="middle">A</forename><surname>Nascimento</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Alexandre</forename><forename type="middle">X</forename><surname>Falcão</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">CIKM &apos;02: Proceedings of the eleventh international conference on Information and knowledge management</title>
				<meeting><address><addrLine>McLean, Virginia, USA</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2002">2002</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
