<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Chemnitz at CLEF 2009 Ad-Hoc TEL Task: Combining Different Retrieval Models and Addressing the Multilinguality</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author role="corresp">
							<persName><forename type="first">Jens</forename><surname>Kürsten</surname></persName>
							<email>jens.kuersten@cs.tu-chemnitz.de</email>
							<affiliation key="aff0">
								<orgName type="department" key="dep1">Faculty of Computer Science</orgName>
								<orgName type="department" key="dep2">Dept. Computer Science and Media</orgName>
								<orgName type="institution">Chemnitz University of Technology</orgName>
								<address>
									<postCode>09107</postCode>
									<settlement>Chemnitz</settlement>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Chemnitz at CLEF 2009 Ad-Hoc TEL Task: Combining Different Retrieval Models and Addressing the Multilinguality</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">EC08B83172FE5F93B790784100EB4E65</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T23:03+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>H.3 [Information Storage and Retrieval]: H.3.1 Content Analysis and Indexing; H.3.3 Information Search and Retrieval Evaluation</term>
					<term>Experimentation</term>
					<term>Cross-Language Information Retrieval</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>In this paper we report our efforts for the participation in the CLEF 2009 Ad-Hoc TEL task. In our second participation we were able to test and evaluate a new feature of the Xtrieval framework, which was the accessibility of the three core retrieval engines Lucene, Lemur and Terrier. This year we submitted 24 experiments in total, 12 each for the monolingual and bilingual subtasks. We compared our baseline experiments to combined runs, where we used two different retrieval models, namely the vector space model (VSM) used in Lucene and the Bose-Einstein model for randomness (BB2) available in the Terrier framework. We found that an almost constant improvement in terms of mean average precision over all provided collections is achievable. Furthermore we tried to benefit from the multilingual contents of the collections by running combined multilingual experiments for both subtasks. The evaluation showed that the used approach achieves small improvements in the monolingual setting of the task. Unfortunately, we were not able to confirm this finding in the bilingual setting, where the multilingual experiments were outperformed by the standard bilingual runs, especially on the English target collection.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction and outline</head><p>The Xtrieval framework <ref type="bibr" target="#b2">[3]</ref>, <ref type="bibr" target="#b1">[2]</ref> was used to prepare and run this year's retrieval experiments in the Ad-Hoc track TEL setting. The core retrieval functionality was provided by Lucene<ref type="foot" target="#foot_0">1</ref> and the Terrier framework <ref type="bibr" target="#b3">[4]</ref>. For the TEL task three different multilingual corpora with content mainly in German, English and French were provided by The European Library. Each collection consists of approximately one million library records. These library records only contain sparse information and have descriptions in multiple languages.</p><p>We conducted monolingual experiments on each of the collections and also submitted experiments for the bilingual task. For the translation of the topics the Google AJAX language API<ref type="foot" target="#foot_1">2</ref> was accessed through a JSON <ref type="foot" target="#foot_2">3</ref> programming interface.</p><p>The remainder of the paper is organized as follows. Section 2 describes the general setup of our system. The individual configurations and the results of our submitted experiments are presented in section 3. In sections 4 and 5 we summarize the results and conclude our observations.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Experimental setup</head><p>This year we were able to choose from various retrieval models and combine the results in the retrieval stage by applying our implementation of the Z-Score operator <ref type="bibr" target="#b4">[5]</ref>. We also used a standard top-k pseudo-relevance feedback algorithm in the retrieval stage, where the values for the top most frequent terms that were obtained from the top documents differed according to the language and used retrieval model. We used the vector space model (VSM) shipped with Lucene and the Bose-Einstein model for randomness (BB2) available in the Terrier framework. We submitted two monolingual baseline runs for all provided collections. Additionally we submitted one monolingual merged experiment and another one in which we tried to benefit from the multilingual character of the collections. The merged monolingual experiments for each collection formed the baseline for two bilingual experiments, where the topics were translated from two different source languages to the corresponding target collection. For two additional bilingual experiments on each target collection we also tried to access the multilingual content of the collections.</p><p>We submitted 9 experiments in which we tried to benefit from the multilingual character of the collections. Therefore we created multiple indexes for each target collection using appropriate stemming and stopword removal for the four most frequent languages. During the retrieval we queried these four indexes and combined the results into one final result list. We needed to translate the topics for all those experiments to the according language of the index, which makes those experiments somewhat multilingual. In table <ref type="table" target="#tab_0">1</ref> we denote the experiments that had multilingual character and present the boost values for the combination in the multilingual result set for each of the experiments in column 'IDs'. These values were chosen according to the occurrence frequency of the language in the corresponding target collection. All runs in the column 'IDs' correspond to an experiment in column 'refer ID' and are directly comparable to this experiment, because we used identical system configurations except for the translation component and the multilingual indexes. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Configurations and Results</head><p>The detailed setup of our experiments and their evaluation results are presented in the following subsections.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">Monolingual Experiments</head><p>We submitted 12 monolingual experiments in total, whereof 4 were submitted for each target collection in German, English and French. For all experiments a language-specific stopword list was applied <ref type="foot" target="#foot_3">4</ref> . We used the stemmers from Snowball<ref type="foot" target="#foot_4">5</ref> for English and French and applied a special n-gram stemmer<ref type="foot" target="#foot_5">6</ref> for German. In table <ref type="table" target="#tab_1">2</ref> the retrieval performance of our experiments is reported in terms of mean average precision (MAP) and the absolute rank of the experiment in the evaluation. We compare the two baseline runs to one combined experiment per target collection. Furthermore we compare the performance of the first baseline run per collection (cut1, cut9, cut17) to the corresponding multilingual experiment (cut4++, cut12++, cut20++). The evaluation of our experiments allows to draw some interesting conclusions. First the overall performance in terms of MAP on the German and French collection were quite similar, while the experiments on the English collection achieved much better results. Interestingly this seemed not to be a flaw in our configuration since we achieved identical position in the ranking over all submitted experiments. Another important observation was that our combined experiments (where different retrieval models were used) always performed better than the baseline run on each of the target collections. However the overall gain was not very large. Furthermore one can conclude that our multilingual approach also worked consistently well by slightly improving MAP (compare cut1 to cut4++, cut9 to cut12++ and cut17 to cut20++).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2">Cross-lingual Experiments</head><p>We submitted 12 experiments for the bilingual subtask, whereof 4 were submitted for each target collection. Two experiments per target collection correspond to the combined monolingual run on that collection. Though two different source topic languages were translated in those experiments. The remaining two runs per target collection had again multilingual character. We translated the topic from the source language to the four most common languages in the target collections, queried the four indexes and combined the results in a multilingual result set. Again the general configuration was equal to the corresponding monolingual reference run for comparability. In table <ref type="table" target="#tab_2">3</ref> we report the evaluation results for each of the bilingual experiments in terms of MAP and the rank over all submitted experiments. Additionally we report our best monolingual experiment for each target collection as baseline for comparison. The evaluation results of our bilingual experiments were very strong. The retrieval performance of our best monolingual runs compared to our best bilingual runs decreased only about 0.6% on the English collection, about 1% on the French collection and about 7,5% on the German collection.</p><p>We still contribute those results to the quality of the Google translation service. Another finding was that the experiments in which we tried to benefit from the multilinguality of the collections also performed quite good in the bilingual setting. In fact one of those experiments performed best on the French collection and on the German collection it performed almost as good as the best experiment. Only on the English collection we could not benefit from the multilinguality, where those two experiments were clearly outperformed by the standard bilingual runs.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Result Analysis -Summary</head><p>The following list provides a summary of the analysis of our retrieval experiments for the Ad-Hoc TEL task at CLEF 2009:</p><p>• Combining retrieval models: Our experiments showed that combining different retrieval models results in a small but consistent gain in terms of MAP over all target collections.</p><p>• Monolingual task: The submitted monolingual experiments achieved strong performance on all target collections. Interestingly the MAP on the French and German collection is almost the same, while the performance is much better on the English collection</p><p>• Bilingual task: Probably due to the used translation service our bilingual experiments performed very good and achieved top results on each target collection. The gap to our best corresponding monolingual runs ranged between 0.6% and 7.5%.</p><p>• Addressing the multilinguality of the collections: We experimented with multilingual configurations and compared them to a baseline experiment. We found that our approach to combine multiple indexed collections works quite good except for the bilingual configurations on the English target collection.</p><p>In our second participation in the CLEF Ad-Hoc TEL task we were able to chose from a wide selection of retrieval models. The Xtrieval framework supports three different retrieval cores now, namely Lucene, Lemur and Terrier. By combining results from Lucene and Terrier we achieved constant gains in terms of mean average precision on all collections over our baseline runs. Again we found that the translation service provided by Google seems to be extremely superior to any other approach or system. We used this service for translating our bilingual and multilingual experiments and got very strong retrieval performance for all those runs. In the future we will further investigate the numerous retrieval models and try to help to develop an open-source retrieval framework for information retrieval evaluation as it was proposed by Ferro and Harman <ref type="bibr" target="#b0">[1]</ref>.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1 :</head><label>1</label><figDesc>Settings for Multilingual Experiments</figDesc><table><row><cell>IDs</cell><cell cols="4">refer ID Collection Topic Translation Boosting</cell></row><row><cell>cut4++, cut7++, cut8++</cell><cell>cut1</cell><cell cols="2">TELONB X2DE</cell><cell>0.568536</cell></row><row><cell></cell><cell></cell><cell cols="2">TELONB X2EN</cell><cell>0.098114</cell></row><row><cell></cell><cell></cell><cell cols="2">TELONB X2FR</cell><cell>0.023721</cell></row><row><cell></cell><cell></cell><cell cols="2">TELONB X2IT</cell><cell>0.014716</cell></row><row><cell cols="2">cut12++, cut15++, cut16++ cut9</cell><cell>TELBL</cell><cell>X2EN</cell><cell>0.595822</cell></row><row><cell></cell><cell></cell><cell>TELBL</cell><cell>X2FR</cell><cell>0.088845</cell></row><row><cell></cell><cell></cell><cell>TELBL</cell><cell>X2DE</cell><cell>0.052568</cell></row><row><cell></cell><cell></cell><cell>TELBL</cell><cell>X2ES</cell><cell>0.028257</cell></row><row><cell cols="2">cut20++, cut23++, cut24++ cut17</cell><cell>TELBNF</cell><cell>X2FR</cell><cell>0.567183</cell></row><row><cell></cell><cell></cell><cell>TELBNF</cell><cell>X2EN</cell><cell>0.118061</cell></row><row><cell></cell><cell></cell><cell>TELBNF</cell><cell>X2DE</cell><cell>0.038291</cell></row><row><cell></cell><cell></cell><cell>TELBNF</cell><cell>X2IT</cell><cell>0.020950</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2 :</head><label>2</label><figDesc>Experimental Results for the Monolingual Task</figDesc><table><row><cell>ID</cell><cell>IR Core / Model</cell><cell>QE</cell><cell>Lang</cell><cell cols="2">MAP Rank</cell></row><row><cell>cut1</cell><cell>Terrier / BB2</cell><cell cols="2">10 / 70 MONO-DE</cell><cell cols="2">0.2602 10/35</cell></row><row><cell>cut2</cell><cell>Lucene / VSM</cell><cell cols="2">10 / 70 MONO-DE</cell><cell>0.2641</cell><cell>9/35</cell></row><row><cell>cut3</cell><cell>merged cut1 &amp; cut2</cell><cell cols="3">10 / 70 MONO-DE 0.2789</cell><cell>2/35</cell></row><row><cell>cut4++</cell><cell>Terrier / BB2</cell><cell cols="2">10 / 70 MONO-DE</cell><cell>0.2713</cell><cell>5/35</cell></row><row><cell>cut9</cell><cell>Terrier / BB2</cell><cell>3 / 11</cell><cell>MONO-EN</cell><cell>0.3864</cell><cell>9/46</cell></row><row><cell>cut10</cell><cell>Lucene / VSM</cell><cell>10 / 7</cell><cell>MONO-EN</cell><cell cols="2">0.3672 15/46</cell></row><row><cell>cut11</cell><cell>merged cut9 &amp; cut10</cell><cell>both</cell><cell cols="2">MONO-EN 0.4071</cell><cell>2/46</cell></row><row><cell cols="2">cut12++ Terrier / BB2</cell><cell>3 / 11</cell><cell>MONO-EN</cell><cell>0.3914</cell><cell>7/46</cell></row><row><cell>cut17</cell><cell>Terrier / BB2</cell><cell>0 / 0</cell><cell>MONO-FR</cell><cell>0.2470</cell><cell>9/35</cell></row><row><cell>cut18</cell><cell>Lucene / VSM</cell><cell>0 / 0</cell><cell>MONO-FR</cell><cell cols="2">0.2399 12/35</cell></row><row><cell>cut19</cell><cell cols="2">merged cut17 &amp; cut18 0 / 0</cell><cell cols="2">MONO-FR 0.2583</cell><cell>2/35</cell></row><row><cell cols="2">cut20++ Terrier / BB2</cell><cell>0 / 0</cell><cell>MONO-FR</cell><cell>0.2540</cell><cell>3/35</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 3 :</head><label>3</label><figDesc>Experimental Results for the Bilingual Task</figDesc><table><row><cell>ID</cell><cell>Model</cell><cell>QE</cell><cell>Lang</cell><cell cols="2">MAP Rank</cell></row><row><cell>cut3</cell><cell>merged cut1 &amp; cut2</cell><cell>10 / 70</cell><cell>MONO-DE</cell><cell>0.2789</cell><cell>2/35</cell></row><row><cell>cut5</cell><cell cols="2">Lucene/VSM &amp; Terrier/BB2 10 / 70</cell><cell cols="2">BILI-EN2DE 0.2583</cell><cell>1/26</cell></row><row><cell>cut6</cell><cell cols="2">Lucene/VSM &amp; Terrier/BB2 10 / 70</cell><cell>BILI-FR2DE</cell><cell>0.2552</cell><cell>3/26</cell></row><row><cell>cut7++</cell><cell>Terrier/BB2</cell><cell>10 / 70</cell><cell>BILI-EN2DE</cell><cell>0.2580</cell><cell>2/26</cell></row><row><cell>cut8++</cell><cell>Terrier/BB2</cell><cell>10 / 70</cell><cell>BILI-FR2DE</cell><cell>0.2444</cell><cell>4/26</cell></row><row><cell>cut11</cell><cell>merged cut9 &amp; cut10</cell><cell cols="2">10 / 7 &amp; 3 / 11 MONO-EN</cell><cell>0.4071</cell><cell>2/46</cell></row><row><cell>cut13</cell><cell cols="4">Lucene/VSM &amp; Terrier/BB2 10 / 7 &amp; 3 / 11 BILI-DE2EN 0.4046</cell><cell>1/43</cell></row><row><cell>cut14</cell><cell cols="3">Lucene/VSM &amp; Terrier/BB2 10 / 7 &amp; 3 / 11 BILI-FR2EN</cell><cell>0.4029</cell><cell>2/43</cell></row><row><cell cols="2">cut15++ Terrier/BB2</cell><cell>3 / 11</cell><cell>BILI-DE2EN</cell><cell>0.3427</cell><cell>9/43</cell></row><row><cell cols="2">cut16++ Terrier/BB2</cell><cell>3 / 11</cell><cell>BILI-FR2EN</cell><cell cols="2">0.3332 11/43</cell></row><row><cell>cut19</cell><cell>merged cut17 &amp; cut18</cell><cell>0 / 0</cell><cell>MONO-FR</cell><cell>0.2583</cell><cell>2/35</cell></row><row><cell>cut21</cell><cell cols="2">Lucene/VSM + Terrier/BB2 0 / 0</cell><cell>BILI-EN2FR</cell><cell>0.2320</cell><cell>4/26</cell></row><row><cell>cut22</cell><cell cols="2">Lucene/VSM + Terrier/BB2 0 / 0</cell><cell>BILI-DE2FR</cell><cell>0.2119</cell><cell>9/26</cell></row><row><cell cols="2">cut23++ Terrier/BB2</cell><cell>0 / 0</cell><cell>BILI-EN2FR</cell><cell>0.2255</cell><cell>5/26</cell></row><row><cell cols="2">cut24++ Terrier/BB2</cell><cell>0 / 0</cell><cell cols="2">BILI-DE2FR 0.2557</cell><cell>1/26</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">http://lucene.apache.org</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_1">http://code.google.com/apis/ajaxlanguage/documentation</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_2">http://json.org</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_3">http://members.unine.ch/jacques.savoy/clef/index.html</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_4">http://snowball.tartarus.org</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="6" xml:id="foot_5">http://www-user.tu-chemnitz.de/ wags/cv/clr.pdf</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgments</head><p>We would like to thank Jaques Savoy and his co-workers for providing numerous resources for language processing. Also, we would like to thank Giorgio M. di Nunzio and Nicola Ferro for developing and operating the DIRECT system 7 .</p><p>This work was partially accomplished in conjunction with the project sachsMedia, which is funded by the Entrepreneurial Regions 8 program of the German Federal Ministry of Education and Research.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Dealing with multilingual information access: Grid experiments at trebleclef</title>
		<author>
			<persName><forename type="first">Nicola</forename><surname>Ferro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Donna</forename><surname>Harman</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Post-proceedings of the Fourth Italian Research Conference on Digital Library Systems</title>
				<meeting><address><addrLine>IRCDL</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2008">2008. 2008</date>
			<biblScope unit="page" from="29" to="32" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Extensible retrieval and evaluation framework: Xtrieval</title>
		<author>
			<persName><forename type="first">Jens</forename><surname>Kürsten</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Thomas</forename><surname>Wilhelm</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Maximilian</forename><surname>Eibl</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">LWA 2008: Lernen -Wissen -Adaption</title>
				<meeting><address><addrLine>Würzburg</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2008-10">October 2008. October 2008</date>
		</imprint>
	</monogr>
	<note>Workshop Proceedings</note>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">The xtrieval framework at clef 2007: Domain-specific track</title>
		<author>
			<persName><forename type="first">Jens</forename><surname>Kürsten</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Thomas</forename><surname>Wilhelm</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Maximilian</forename><surname>Eibl</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">LNCS -Advances in Multilingual and Multimodal Information Retrieval</title>
				<editor>
			<persName><forename type="first">C</forename><surname>Peters</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">V</forename><surname>Jijkoun</surname></persName>
		</editor>
		<editor>
			<persName><surname>Th</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">H</forename><surname>Mandl</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">D</forename><forename type="middle">W</forename><surname>Müller</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Oard</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">V</forename><surname>Peñas</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">D</forename><surname>Petras</surname></persName>
		</editor>
		<editor>
			<persName><surname>Santos</surname></persName>
		</editor>
		<meeting><address><addrLine>Berlin</addrLine></address></meeting>
		<imprint>
			<publisher>Springer Verlag</publisher>
			<date type="published" when="2008">2008</date>
			<biblScope unit="volume">5152</biblScope>
			<biblScope unit="page" from="174" to="181" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Research directions in terrier: a search engine for advanced retrieval on the web</title>
		<author>
			<persName><forename type="first">Iadh</forename><surname>Ounis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Christina</forename><surname>Lioma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Craig</forename><surname>Macdonald</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Vassilis</forename><surname>Plachouras</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Novatica/UPGRADE Special Issue on Next Generation Web Search</title>
				<imprint>
			<date type="published" when="2007">2007</date>
			<biblScope unit="page" from="49" to="56" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Data fusion for effective european monolingual information retrieval</title>
		<author>
			<persName><forename type="first">Jaques</forename><surname>Savoy</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Working Notes for the CLEF 2004 Workshop</title>
				<meeting><address><addrLine>Bath, UK</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2004-09">September 2004</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<monogr>
		<title level="m" type="main">The Innovation Initiative for the New German Federal States</title>
		<imprint/>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
