<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Twenty-One at CLEF-2000: Translation resources, merging strategies and relevance feedback</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Djoerd</forename><surname>Hiemstra</surname></persName>
							<email>hiemstra@cs.utwente.nl</email>
							<affiliation key="aff1">
								<orgName type="laboratory">TNO-TPD</orgName>
								<address>
									<postBox>P.O. Box 155</postBox>
									<postCode>2600 AD</postCode>
									<settlement>Delft</settlement>
									<country key="NL">The Netherlands</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Wessel</forename><surname>Kraaij</surname></persName>
							<email>kraaij@tpd.tno.nl</email>
							<affiliation key="aff1">
								<orgName type="laboratory">TNO-TPD</orgName>
								<address>
									<postBox>P.O. Box 155</postBox>
									<postCode>2600 AD</postCode>
									<settlement>Delft</settlement>
									<country key="NL">The Netherlands</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Renée</forename><surname>Pohlmann</surname></persName>
							<email>pohlmann@tpd.tno.nl</email>
						</author>
						<author>
							<persName><forename type="first">Thijs</forename><surname>Westerveld</surname></persName>
						</author>
						<author>
							<affiliation key="aff0">
								<orgName type="institution">Ý University of Twente</orgName>
								<address>
									<postBox>CTIT, P.O. Box 217</postBox>
									<postCode>7500 AE</postCode>
									<settlement>Enschede</settlement>
									<country key="NL">The Netherlands</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Twenty-One at CLEF-2000: Translation resources, merging strategies and relevance feedback</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">D184EF799AE976D24A6D88DC11CCA6C4</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-25T03:10+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>This paper describes the official runs of the Twenty-One group for CLEF-2000. The Twenty-One group participated in the monolingual, bilingual and multilingual tasks. The following new techniques are introduced in this paper. In the bilingual task we experimented with different methods to estimate translation probabilities. In the multilingual task we experimented with refinements on raw-score merging techniques and with a new relevance feedback algorithm that re-estimates both the model's translation probabilities and the relevance weights. Finally, we performed preliminary experiments to exploit the web to generate translation probabilities and bilingual dictionaries, notably for English-Italian and English-Dutch.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>Twenty-One is a project funded by the EU Telematics Applications programme, sector Information Engineering. The project subtitle is "Development of a Multimedia Information Transaction and Dissemination Tool". Twenty-One started early 1996 and was completed in June 1999. Because the TREC ad-hoc and cross-language information retrieval (CLIR) tasks fitted our needs to evaluate the system on the aspects of monolingual and cross-language retrieval performance, TNO-TPD and University of Twente participated under the flag of "Twenty-One" in TREC-6 / 7 / 8. Since the cooperation is continued in other projects: Olive and Druid, we have decided to continue our participation in CLEF as "Twenty-One". 1 For all tasks, we used the TNO vector retrieval engine. The engine supports several term weighting schemes. The principal term weighting scheme we used is the "linguistically motivated probabilistic model of information retrieval" <ref type="bibr" target="#b1">[2,</ref><ref type="bibr" target="#b3">4]</ref> explained below.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">The approach</head><p>All runs were carried out with an information retrieval system based on a simple unigram language model. The basic idea is that documents can be represented by simple statistical language models. Now, if a query is more probable given a language model based on document ½ , than given e.g. a language model based on document ¾ , then we hypothesise that the document ½ is more relevant to the query than document ¾ . Thus the probability of generating a certain query given a document-based language model can serve as a score to rank documents with respect to relevance. 1 Information about Twenty-one, Olive and Druid is available at http://dis.tpd.tno.nl/ Formula 1 shows the basic idea of this approach to information retrieval, where the document-based language model is interpolated with a background language model to compensate for sparseness. In the formula, Ì is a random variable for the query term on position in the query (½ Ò, where Ò is the query length), which sample space is the set Ø ´¼µ Ø ´½µ ¡ ¡ ¡ Ø ´Ñµ of all terms in the collection. The probability measure È ´Ì µ defines the probability of drawing a term at random from the collection, È ´Ì µ defines the probability of drawing a term at random from the document; and defines the importance of each query term. The marginal probability of relevance È ´ µ might be assumed uniformly distributed over the documents in which case it may be ignored in the above formula.</p><formula xml:id="formula_0">È ´Ì½ Ì ¾ ¡ ¡ ¡ Ì Ò µÈ ´ µ È ´ µ Ò ½ ´½ µÈ ´Ì µ • È ´Ì µ (1)</formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1">A model of cross-language information retrieval</head><p>Information retrieval models and statistical translation models can be integrated into one unifying model for cross-language information retrieval <ref type="bibr" target="#b0">[1,</ref><ref type="bibr" target="#b3">4]</ref>. Let Ë be a random variable for the source language query term on position . Each document gets a score defined by the following formula.</p><formula xml:id="formula_1">È ´Ë½ Ë ¾ ¡ ¡ ¡ Ë Ò µ È ´ µ È ´ µ Ò ½ Ñ ½ È ´Ë Ì Ø ´ µ µ´´½ µÈ ´Ì Ø ´ µ µ • È ´Ì Ø ´ µ µµ<label>(2)</label></formula><p>In the formula, the probability measure È ´Ë Ì Ø ´ µ µ defines the translation probabilities.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2">Translation in practice</head><p>In practice, the statistical translation model will be used as follows. The automatic query formulation process will translate the query Ë ½ Ë ¾ ¡ ¡ ¡ Ë Ò using a probabilistic dictionary. The probabilistic dictionary is a dictionary that list pairs ´× Øµ together with their probability of occurrence, where × is from the sample space of Ë and Ø is from the sample space of Ì . For each Ë there will be one or more realisations Ø of Ì for which È ´Ë Ì Ø µ ¼, which will be called the possible translations of Ë . The possible translations should be grouped for each to search the document collection, resulting in a structured query. For instance, suppose the original French query on an English collection is "déchets dangereux", then possible translations of "déchets" might be "waste", "litter" or "garbage", possible translations of "dangereux" might be "dangerous" or "hazardous" and the structured query can be presented as follows.</p><p>´´waste litter garbageµ ´dangerous hazardousµµ</p><p>The product from ½ to Ò (in this case Ò ¾ ) of equation 2 is represented above by using the comma as is done in the representation of a query of length 2 as Ì ½ Ì ¾ . The sum from ½ to Ñ of equation 2 is represented by displaying only the realisations of Ì for which È ´Ë Ì µ ¼ and by separating those by ' '. So, in practice, translation takes place during automatic query formulation (query translation), resulting in a structured query like the one displayed above that is matched against each document in the collection. Unless stated otherwise, whenever this paper mentions 'query terms', it will denote the target language query terms: realisations of Ì . Realisations of Ë , the source language query terms, will usually left implicit. The combination of the structured query representation and the translation probabilities will implicitly define the sequence of the source language query terms Ë ½ Ë ¾ ¡ ¡ ¡ Ë Ò , but the actual realisation of the sequence is not important to the system.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.3">Probability estimation</head><p>The prior probability of relevance È ´ µ, the probability of term occurrence in the collection È ´Ì µ and the probability of term occurrence in the relevant document È ´Ì µ are defined by the collection that is searched. For the evaluations reported in this paper, the following definitions were used, where Ø ´Ø µ denotes the number of occurrences of the term Ø in the document , and ´Øµ denotes the number of documents in which the term Ø occurs. Equation 3 is the definition used for the unofficial "document length normalisation" runs reported in section 5.</p><formula xml:id="formula_2">È ´ µ È Ø Ø ´Ø µ È Ø Ø ´Ø µ (3) È ´Ì Ø µ Ø ´Ø µ È Ø Ø ´Ø µ (4) È ´Ì Ø µ ´Ø µ È Ø ´Øµ<label>(5)</label></formula><p>The translation probabilities È ´Ë Ì µ and the value of , however, are unknown. The collection that is searched was not translated, or if it was translated, the translations are not available. Translation probabilities should therefore be estimated from other data, for instance from a parallel corpus. The value of determines the importance of the source language query term. If ½ then the system will assign zero probability to documents that do not contain any of the possible translations of the original query term on position . In this case, a possible translation of the source language term is mandatory in the retrieved documents. If</p><p>¼ then the possible translations of the original query term on position will not affect the final ranking. In this case, the source language query term is treated as if it were a stop word. For ad-hoc queries, it is not known which of the original query terms are important and which are not important and a constant value for each is taken. The system's default value is ¼ ¿.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.4">Implementation</head><p>Equation 2 is not implemented as is, but instead it is rewritten into a weighting algorithm that assigns zero weight to terms that do not occur in the document. Filling in the definitions of equation 3, 4 and 5 in equation 2 results in the following formula. The probability measure È ´Ë Ì Ø ´ µ µ will be replaced by the translation probability estimates ´ µ.</p><formula xml:id="formula_3">È ´ Ë ½ Ë ¾ ¡ ¡ ¡ Ë Ò µ È Ø Ø ´Ø µ È Ø Ø ´Ø µ Ò ½ Ñ ½ ´ µ´´½ µ ´Ø´ µ µ È Ø ´Øµ • Ø ´Ø´ µ µ È Ø Ø ´Ø µ µ</formula><p>The translation probabilities can be moved into the inner sum. As summing is associative and commutative, it is not necessary to calculate each probability separately before adding them. Instead, respectively the document frequencies and the term frequencies of the disjuncts can be added beforehand, properly multiplied by the translation probabilities. Only in the big sum is constant for every addition and can therefore be moved outside the sum, resulting in:</p><formula xml:id="formula_4">È ´ Ë ½ Ë ¾ ¡ ¡ ¡ Ë Ò µ È Ø Ø ´Ø µ È Ø Ø ´Ø µ Ò ½ ´´½ µ È Ñ ½ ´ µ ´Ø´ µ µ È Ø ´Øµ • È Ñ ½ ´ µØ ´Ø´ µ µ È Ø Ø ´Ø µ µ</formula><p>Using simple calculus (see e.g. <ref type="bibr" target="#b2">[3]</ref>), the probability measure can now be rewritten into a term weighting algorithm that assigns zero weight to non-matching terms, resulting in equation 6. The formula ranks documents in exactly the same order as equation 2.</p><formula xml:id="formula_5">È ´ Ë ½ Ë ¾ ¡ ¡ ¡ Ë Ò µ » ÐÓ ´ÈØ Ø ´Ø µµ • Ò ½ ÐÓ ´½• ´ÈÑ ½ ´ µØ ´Ø´ µ µµ È Ø ´Øµ ´½ µ´È Ñ ½ ´ µ ´Ø´ µ µµ È Ø Ø ´Ø µ µ (<label>6</label></formula><formula xml:id="formula_6">)</formula><p>Equation 6 is the algorithm implemented in the TNO retrieval engine. It contains a weighted sum of respectively the term frequencies and the document frequencies where the weights are determined by the translation probabilities ´ µ. Unweighted summing of frequencies was used before for on-line stemming in <ref type="bibr" target="#b4">[5]</ref> in a vector space model retrieval system.</p><p>The model does not require the translation probabilities ´ µ to sum up to one for each , since they are conditioned on the target language query term and not on the source language query term. Interestingly, for the final ranking it does not matter what the actual sum of the translation probabilities is. Only the relative proportions of the translations define the final ranking of documents. This can be seen by ´ µ which occurs in the numerator and in the denominator of the big fraction in equation 6.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.5">A Relevance feedback method for cross-language retrieval</head><p>This paper introduces a new relevance feedback method for cross-language information retrieval. If there were some known relevant documents, then the values of ´ µ and could be re-estimated from that data. The idea is the following. Suppose there are three known relevant English documents to the French query "déchets dangereux". If two out of three documents contain the term "waste" and none contain the terms "litter" and "garbage" then this is an indication that "waste" is the correct translation and should be assigned a higher translation probability than 'litter" and "garbage". If only one of the three known relevant document contains one or more possible translations of "dangereux" then this is an indication that the original query term "déchets" is more important (possible translations occur in more relevant documents) than the original query term "dangereux" and the value of should be higher for "déchets" than for "dangereux".</p><p>The actual re-estimation of ´ µ and was done by iteratively applying the EM-algorithm defined by the formulas in equation 7. In the algorithm, ´ µ ´Ôµ and ´Ôµ denote the values on the Ôth iteration. The values are initialised with the translation probabilities from the dictionary and with ´</p><formula xml:id="formula_7">µ ´Ô•½µ ½ Ö Ö ½ ´ µ ´Ôµ ´´½ ´Ôµ µÈ ´Ì Ø ´ µ µ • ´Ôµ È ´Ì Ø ´ µ µµ È Ñ Ð ½ ´Ðµ ´Ôµ ´´½ ´Ôµ µÈ ´Ì Ø ´Ðµ µ • ´Ôµ È ´Ì Ø ´Ðµ µµ ´Ô•½µ ½ Ö Ö ½ ´Ôµ ´ÈÑ Ð ½ ´Ðµ ´Ôµ È ´Ì Ø ´Ðµ µµ È Ñ Ð ½ ´Ðµ ´Ôµ ´´½ ´Ôµ µÈ ´Ì Ø ´Ðµ µ • ´Ôµ È ´Ì Ø ´Ðµ µµ<label>(7)</label></formula><p>The re-estimation of ´ µ and was done from 'pseudo-relevant' documents. First the top 10 documents were retrieved using the default values of ´ µ and and then the feedback algorithm was used on these documents to find the new values. The actual algorithm implemented was a variation of equation 7 of the form: ´½ ´Ö • ½ µ µ ¡ ´default value • È Ö ½ µ to avoid that e.g.</p><p>½ after re-estimation.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Translation resources</head><p>As in previous years we applied a dictionary based query translation approach. The translations were based on the VLIS lexical database of Van Dale publishers <ref type="bibr" target="#b1">[2]</ref>. Because VLIS currently lacks translations into Italian, we used two other resources: i) the Systran web based MT engine ii) a probabilistic lexicon based a parallel web corpus. The next section will describe the construction of this new resource in more detail.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">Parallel web corpora</head><p>We developed three parallel corpora based on web pages in close cooperation with RALI, Université de Montréal. RALI already had developed an English-French parallel corpus of web pages, so it seemed interesting to investigate the feasibility of a full multilingual system based on web derived lexical resources only. We used the PTMiner tool <ref type="bibr" target="#b6">[7]</ref> to find web pages which have a high probability to be translations of each other. The mining process consists of the following steps:</p><p>1. Query a web search engine for web pages with a hyperlink anchor text "English version" and respective variants.</p><p>2. (For each web site) Query a web search engine for all web pages on a particular site.</p><p>3. (For each web site) Try to find pairs of path names that match certain patterns, e.g.: /department/tt/english/home.html and /department/tt/italian.html.</p><p>4. (For each pair) download web pages, perform a language check using a probabilistic language classifier, remove pages which are not positively identified as being written in a particular language.</p><p>The mining process was run for three language pairs and resulted in three modest size parallel corpora. Table <ref type="table" target="#tab_0">1</ref> lists sizes of the corpus during intermediate steps. Due to the dynamic nature of the web, a lot of pages that have been indexed, do not exist anymore. Sometimes a site is down for maintenance. Finally a lot of pages are simply place holders for images and are discarded by the language identification step. These parallel corpora have been used in different ways: i) to refine the estimates of translation probabilities of a dictionary based translation system (corpus based probability estimation) ii) to construct simple statistical translation models (IBM model 1) <ref type="bibr" target="#b6">[7]</ref> . The former application will be described in more detail in Section 5.2 the latter in Section 5.3. The translation models for English-Italian and English-German, complemented with an already existing model for English-French formed also the basis for a full corpus based translation multilingual run which is described elsewhere in this volume <ref type="bibr" target="#b5">[6]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Merging intermediate runs</head><p>Our strategy to multilingual retrieval is to translate the query into the document languages, perform separate language specific runs and merge the results into a single result file. In previous CLIR evaluations, we compared different merging strategies: round robin Here the idea is that document scores are not comparable across collections, because we are basically ignorant about the distribution of relevant documents in the retrieved lists, round robin assumes that these distributions are similar across languages.</p><p>raw score This type of merging assumes that document scores are comparable across collections.</p><p>rank based It has been observed that the relationship between probability of relevance and the log of the rank of a document can be approximated by a linear function, at least for a certain class of IR systems. If a training collection is available, one can estimate the parameters of this relationship by applying regression. Merging can subsequently be based on the estimated probability of relevance.</p><p>Note that the actual score of a document is only used to rank documents, but that merging is based on the rank, not on the score.</p><p>The new CLEF multilingual task is based on a new document collection which makes it hard to compute reliable estimates for the linear parameters; a training set is not available. A second disadvantage of the rank based merging strategy is that the linear function generalises across topics. Unfortunately in the multilingual task, the distribution of relevant documents over the subcollections is quite skewed. All collections have several (differing) topics without relevant documents, so applying a rank based merging strategy would hurt the performance for these topics, because the proportion of retrieved documents in every collection is the same for every topic.</p><p>The raw score merging strategy (which proved succesful last year) does not need training data and also does not suffer from the equal proportions strategy. Unfortunately, usually scores are not totally compatible across collections. We have tried to identify factors which cause these differences. We have applied two normalization techniques. First of all we treat term translations as a weighted concept vector (cf. section 2). That means that we can normalise scores across topics by dividing the score by the query length. This amounts to computing the geometric avarage of probabilities per query concept. Secondly, we have observed that collection size has a large influence on the occurence probability estimates È ´Ì µ because the probability of rare terms is inversely proportional to the collection size.  Figure <ref type="figure">4</ref> shows the probability estimates of a sample of words of 1 document when we add more documents to the collection. The occurrence probability of common words stabilises fast when the collection size increases. The more rare a word is however, the higher is the degree of overestimation of its occurrence probability. This effect is a consequence of the sparse data problem. In fact, a small collection will never yield correct term occurrence probability estimates.</p><p>The collection-size dependency of collection-frequency (or global term frequency) estimates has a direct influence on the distribution of document scores for a particular query. When the collection is small, the scores will be lower than the scores on a large collection. This is due to the fact that the score we study is based on the maximum likelihood ratio. So the median of the distribution of document scores for a particular topic (set) is inversely related with the collection size. Thus when we use the raw scores of different subcollections as a basis for merging, large collections will be favoured.</p><p>We hypothesised that we could improve the merging process, if we could correct the estimates for their dependence on the collection size. Suppose we have just two collections with a different size (and different language): ½ , ¾ with vocabulary size Î ½ ,Î ¾ and number of tokens Ì ½ , Ì ¾ respectively, with Ì ½ Ì ¾ . Now we could try to either extrapolate the term occurrence probability estimates on collection ½ to a hypothetical collection with Ì ¾ tokens or try to 'downscale' the term occurrence probability estimates of a term from ¾ to vocabulary size Î ½ .</p><p>The first option seems cumbersome, because we have hardly information to guide the extrapolation process. The second option, trying to adapt the estimates of the large collection to the small collection, seems more viable. The idea is to adapt the probability estimates of rare terms in such a way, that they will become 'compatible' with the estimates on the small collection. As shown in figure <ref type="figure">4</ref> the estimates of frequent terms stabilise soon. Our idea is to construct a mapping function which maps the probability estimates to the small collection domain. The mapping function has the following requirements: a probability ½ Ì ¾ has to be mapped to ½ Ì ½ . So the probability is multiplied by the factor Ì ¾ Ì ½ and probabilities Ô larger than ½ Ì ¾ will be multiplied by a factor which decreases for larger Ô. In fact we only want very small changes for Ô ½¼ ¿ . A function which meets these properties is the polynomial ´Üµ Ü Ü ¾ (where Ü ÐÓ ´Ôµ and</p><formula xml:id="formula_8">Ì¾ Ì½ Ì ¾ ¾ µ.</formula><p>Because we have re-estimated the probabilities, one would expect that the probabilities have to be re-normalised ( Ô¼´Ø µ Ô´Ø µ È Î¾ Ô´Ø µ ). However, this has the result that all global probabilities (also those of relatively frequent words) are increased, which will increase the score of all documents, i.e. will have the opposite effect of what we want. So we decide not to re-normalise, because a smaller corpus would also have a smaller vocabulary, which would compensate for the increase in probability mass which is a result of the transformation.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Results</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.1">Monolingual runs</head><p>We indexed the collections in the 4 languages separately. All documents were lemmatised using the Xelda morphological toolkit from Xerox XRCE and stopped with language specific stoplists. For German, we splitted compounds and added both the full compound and its parts to the index. This strategy is motivated by our experience with a Dutch corpus (Dutch is also a compounding language) <ref type="bibr" target="#b7">[8]</ref> and tests on the TREC CLIR test collection. Table <ref type="table" target="#tab_2">2</ref> shows the results of the monolingual runs, runs in bold are judged runs, runs in italic font are unofficial runs (mostly post-hoc). The table also lists the proportion of documents which has been judged. The standard runs include fuzzy lookup of unknown words. The expand option adds close orthographical variants for every query term. The official runs were done without document length normalisation defined by equation 3.</p><p>run name avp above median description % j@1000 %j@100 %j@10 tnoutdd1 0. The first thing that strikes us, is that the pool depth is 50, contrary to what has been practice in TREC in which the top 100 documents are judged for relevance. Section 5.4 analyses the CLEF collection further. Length normalisation usually gives a modest improvement in average precision. The 'expand' option was especially effective for German. The reason is probably that compound parts are not always properly lemmatised by the German morphology. Especially the German run performs well with 28 out of 37 topics above average. This relatively good performance is probably due to the morphology, which includes compound splitting.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.2">Bilingual runs</head><p>Table <ref type="table" target="#tab_3">3</ref> lists the results of the bilingual runs. All runs use Dutch as a query language. The base run of 0.3069 can be improved by several techniques: a higher lambda, document length normalisation or Porter stemming instead of dictionary based stemming. The latter can be explained by the fact that Porter's algorithm is an aggressive stemmer that also removes most of the derivational affixes. This is usually beneficial to retrieval performance. The experiment with corpus based frequencies yielded disappointing results. We first generated topic translations in a standard fashion based on VLIS. Subsequently we replaced the translation probabilities È ´ÛAE Ä Û AE µ by rough corpus based estimates. We simply looked up all English sentences which contained the translation and determined the proportion of the corresponding (aligned) Dutch sentences that contained the original Dutch query word. If the pair was not found, the original probability was left unchanged. Unfortunately a lot of the query terms and translations were not found in the aligned corpus, because they were lemmatised whereas the corpus was not lemmatised. At least this mismatch did hurt the estimates. The procedure resulted in high translation probabilities for words that did not occur in the corpus and low probabilities for words that did occur. The pseudo relevance feedback runs were done with the experimental language models retrieval engine at the University of Twente, using an index based on the Porter stemming algorithm. The run tagged with tnoutne3-stem is the baseline run for this system. The official pseudo relevance feedback run used the top 10 documents retrieved to re-estimate relevance weights and translation probabilities, but turned out to contain a bug. The unofficial fixed run tnoutne4-fix performs a little bit worse than the baseline. The run tnoutne4-retro uses the relevant documents to re-estimate the probabilities retrospectively (see e.g. <ref type="bibr" target="#b8">[9]</ref>). This run reaches an impressive performance of 0.4695 average precision, much higher even than the best monolingual English run. This indicates that the algorithm might be helpful in an interactive setting where the user's feedback is used to retrieve a new, improved, set of documents. Apparently, the top 10 retrieved contains too much noise to be useful for the re-estimation of the model's parameters.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.3">Multilingual runs</head><p>Table <ref type="table" target="#tab_4">4</ref> shows that our best multilingual run was a run with Dutch as a query language. This is on one hand surprising (because this run is composed of 4 bilingual runs instead of 3 for the EN X run. But the translation is based on the VLIS lexical database which is built on lexical relations with Dutch as a source language. Thus the translations in the NL X case are much cleaner than the EN X case. In the latter case, Dutch serves as a pivot language. On the other hand, the NL IT translation is quite cumbersome. We first used Xelda to translate the Dutch queries to English stopped and lemmatised files. These files were subsequently translated by Systran.</p><p>Another interesting point is that the intermediate bilingual run based on the parallel web corpus performed quite well, with an average precision of 0.2750 versus 0.3203 of Systran. The translation of this run is based on a translation model trained on the parallel web corpus. The English topics were simply stopped and translated by the translation model. We took the most probable translation and used that as Italian query. We plan to experiment with a more refined approach where we import the translation probabilities into structured queries. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.4">The CLEF collection</head><p>This section reports on some of the statistics of the CLEF collection and compares it to the TREC crosslanguage collection. Table <ref type="table" target="#tab_6">6</ref> lists the same information for the TREC collection. The collections are actually quite different. First of all, the CLEF collection is almost half the size of the TREC collection and heavily biased towards German and English documents. Although the CLEF organisation decided to judge only the top 50 of documents retrieved and not the top 100 documents retrieved as in TREC, the number of documents judged per topic is only a little lower for the CLEF collection: about 814 documents per topic vs. 834 for TREC. Given the fact that the 56 TREC topics were developed over a period of two years and the CLEF collection has 40 topics already, the organisation actually did more work this year compared to pervious years. Another striking difference is the number of relevant documents per topic, only 57 for CLEF and 116 for TREC. This might actually make the decision to only judge the top 50 of runs not that harmful for the usefulness of the CLEF evaluation results.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6">Conclusions</head><p>This year's evaluation has confirmed that cross-language retrieval based on structured queries, no matter what the translation resources are, is a powerful technique. Re-estimating model parameters based on pseudo relevant documents does not result in improvement of retrieval performance. However, the relevance weighting algorithm shows an impressive performance gain if the relevant documents are used retrospectively. This indicates that the algorithm might in fact be a valuable tool for processing user feedback in an inter-active setting. Finally, merging based on the collection size re-estimation technique proved not successful. Further analysis is needed why the technique did not work on this collection, as it was quite successful on the TREC-8 collection.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head></head><label></label><figDesc>´¼µ ¼ ¿. The reestimation formulas should be used simultaneously for each Ô until the values do not change significantly anymore.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Probability estimates vs collection size</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1 :</head><label>1</label><figDesc>Intermediate sizes during corpus construction</figDesc><table><row><cell cols="5">language nr of web sites nr of candidate pages nr of candidate pairs retrieved + cleaned pairs</cell></row><row><cell>EN-IT</cell><cell>3651</cell><cell>1053649</cell><cell>23447</cell><cell>4768</cell></row><row><cell>EN-DE</cell><cell>3817</cell><cell>1828906</cell><cell>33577</cell><cell>5743</cell></row><row><cell>EN-NL</cell><cell>3004</cell><cell>1170082</cell><cell>24738</cell><cell>2907</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 2 :</head><label>2</label><figDesc>Results of the monolingual runs</figDesc><table><row><cell></cell><cell>3760</cell><cell></cell><cell>standard</cell><cell>18.64</cell><cell>79.05</cell><cell>100</cell></row><row><cell>tnoutdd2</cell><cell>0.3961</cell><cell>28/37</cell><cell>+expand</cell><cell>18.72</cell><cell>81.22</cell><cell>100</cell></row><row><cell>tnoutdd2l</cell><cell>0.3968</cell><cell></cell><cell>+length normalisation</cell><cell>18.58</cell><cell>78.22</cell><cell>97.50</cell></row><row><cell>tnoutff1</cell><cell>0.4551</cell><cell></cell><cell>standard</cell><cell>16.13</cell><cell>79.42</cell><cell>100</cell></row><row><cell>tnoutff2</cell><cell>0.4471</cell><cell>18/34</cell><cell>+expand</cell><cell>16.21</cell><cell>80.88</cell><cell>100</cell></row><row><cell>tnoutff2l</cell><cell>0.4529</cell><cell></cell><cell>+length normalisation</cell><cell>16.00</cell><cell>77.88</cell><cell>97.50</cell></row><row><cell>tnoutii1</cell><cell>0.4677</cell><cell></cell><cell>standard</cell><cell>16.59</cell><cell>78.92</cell><cell>100</cell></row><row><cell>tnoutii2</cell><cell>0.4709</cell><cell>18/34</cell><cell>+expand</cell><cell>16.67</cell><cell>80.33</cell><cell>100</cell></row><row><cell>tnoutii2l</cell><cell>0.4808</cell><cell></cell><cell>+length normalisation</cell><cell>16.66</cell><cell>77.25</cell><cell>98</cell></row><row><cell cols="2">tnoutee01i 0.4200</cell><cell></cell><cell>standard</cell><cell>17.81</cell><cell>71.10</cell><cell>100</cell></row><row><cell cols="2">tnoutee01 0.4169</cell><cell></cell><cell>+expand</cell><cell>17.84</cell><cell>70.75</cell><cell>99.75</cell></row><row><cell cols="2">tnoutee01l 0.4273</cell><cell></cell><cell>+length normalisation</cell><cell>17.82</cell><cell>69.30</cell><cell>98.00</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 3 :</head><label>3</label><figDesc>Results of the bilingual runs</figDesc><table><row><cell>run name</cell><cell>avp</cell><cell>above median</cell><cell></cell><cell>description</cell></row><row><cell>tnoutne1</cell><cell>0.3069</cell><cell>27/33</cell><cell cols="2">standard</cell></row><row><cell>tnoutne1l</cell><cell>0.3278</cell><cell>-</cell><cell cols="2">+ doclen norm</cell></row><row><cell>tnoutne1p</cell><cell>0.3442</cell><cell>-</cell><cell>+</cell><cell>¼</cell></row><row><cell>tnoutne2</cell><cell>0.2762</cell><cell>25/33</cell><cell cols="2">corpus frequencies</cell></row><row><cell cols="2">tnoutne3-stem 0.3366</cell><cell>-</cell><cell cols="2">Porter stemmer +doclen norm</cell></row><row><cell>tnoutne4</cell><cell>0.2946</cell><cell>20/33</cell><cell cols="2">pseudo relevance feedback (PRF)</cell></row><row><cell>tnoutne4-fix</cell><cell>0.3266</cell><cell>-</cell><cell cols="2">PRF bugfix +doclen norm, Porter</cell></row><row><cell cols="2">tnoutne4-retro 0.4695</cell><cell>-</cell><cell cols="2">retrospective relevance feedback</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_4"><head>Table 4 :</head><label>4</label><figDesc>Results of the</figDesc><table><row><cell>run name</cell><cell>avp</cell><cell>above median</cell><cell cols="2">description</cell></row><row><cell>tnoutex1</cell><cell>0.2214</cell><cell>25/40</cell><cell cols="2">baseline run</cell></row><row><cell cols="2">tnoutex2 0.2165</cell><cell>26/40</cell><cell>merged</cell></row><row><cell cols="2">tnoutex2f 0.2219</cell><cell></cell><cell>fixed</cell></row><row><cell>tnoutex3</cell><cell>0.1960</cell><cell>25/40</cell><cell cols="2">Web based EN-IT lexicon</cell></row><row><cell cols="2">tnoutnx1 0.2256</cell><cell>23/40</cell><cell cols="2">query language is Dutch</cell></row><row><cell></cell><cell></cell><cell></cell><cell>AE Ê</cell><cell>Á Ì runs</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_5"><head>Table 5 :</head><label>5</label><figDesc>Table 5  lists the size, number of judged documents, number of relevant documents and the judged fraction, which is the part of the collection that is judged per topic. CLEF collection statistics, 40 topics </figDesc><table><row><cell>collection</cell><cell cols="3">total judged relevant no hits</cell><cell>judged</cell></row><row><cell></cell><cell>docs.</cell><cell>docs.</cell><cell>docs. in topic</cell><cell>fraction</cell></row><row><cell>english</cell><cell cols="2">110,250 14,737</cell><cell cols="2">579 2, 6, 8, 23, 25, 27, 35 0.0033</cell></row><row><cell>french</cell><cell>44,013</cell><cell>8,434</cell><cell>528 2, 4, 14, 27, 28, 36</cell><cell>0.0048</cell></row><row><cell>german</cell><cell cols="2">153,694 12,283</cell><cell>821 2, 28, 36</cell><cell>0.0020</cell></row><row><cell>italian</cell><cell>58,051</cell><cell>8,112</cell><cell>338 3, 6, 14, 27, 28, 40</cell><cell>0.0035</cell></row><row><cell>total</cell><cell cols="2">366,008 43,566</cell><cell>2,266</cell><cell>0.0022</cell></row><row><cell>collection</cell><cell cols="3">total judged relevant no hits</cell><cell>judged</cell></row><row><cell></cell><cell>docs.</cell><cell>docs.</cell><cell>docs. in topic</cell><cell>fraction</cell></row><row><cell>english</cell><cell cols="2">242,866 18,783</cell><cell>2,645 26, 46, 59, 63, 66, 75</cell><cell>0.0014</cell></row><row><cell>french</cell><cell cols="2">141,637 11,881</cell><cell>1,569 76</cell><cell>0.0015</cell></row><row><cell>german</cell><cell>185,099</cell><cell>8,656</cell><cell>1,634 26, 60 ,75, 76</cell><cell>0.0008</cell></row><row><cell>italian</cell><cell>62,359</cell><cell>7,396</cell><cell cols="2">671 26, 44, 51, 60, 63, 75, 80 0.0021</cell></row><row><cell>total</cell><cell cols="2">631,961 46,716</cell><cell>6,519</cell><cell>0.0013</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_6"><head>Table 6 :</head><label>6</label><figDesc>TREC collection statistics, 56 topics (26-81)</figDesc><table /></figure>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgements</head><p>We would like to thank the DRUID project for sponsoring the translation of the topic set into Dutch. We thank Xerox XRCE for making the Xelda morphological toolkit available to us. Furthermore we would like to thank Jiang Chen (RALI, Université de Montréal), Jian-Yun Nie (RALI) for help with the PTMiner web mining tools, and Michel Simard (RALI) for helping with the construction of aligned corpora and building translation models.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Disambiguation strategies for cross-language information retrieval</title>
		<author>
			<persName><forename type="first">D</forename><surname>Hiemstra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">M G</forename><surname>De</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jong</forename></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the third European Conference on Research and Advanced Technology for Digital Libraries</title>
				<meeting>the third European Conference on Research and Advanced Technology for Digital Libraries</meeting>
		<imprint>
			<date type="published" when="1999">1999</date>
			<biblScope unit="page" from="274" to="293" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Twenty-One at TREC-7: Ad-hoc and cross-language track</title>
		<author>
			<persName><forename type="first">D</forename><surname>Hiemstra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Kraaij</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the seventh Text Retrieval Conference TREC-7, NIST Special Publication</title>
				<meeting>the seventh Text Retrieval Conference TREC-7, NIST Special Publication</meeting>
		<imprint>
			<date type="published" when="1999">1999</date>
			<biblScope unit="page" from="227" to="238" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">A probabilistic justification for using tf.idf term weighting in information retrieval</title>
		<author>
			<persName><forename type="first">D</forename><surname>Hiemstra</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">ternational Journal on Digital Libraries</title>
		<imprint>
			<biblScope unit="volume">3</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page" from="131" to="139" />
			<date type="published" when="2000">2000</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Twenty-one at TREC-8: using language technology for information retrieval</title>
		<author>
			<persName><forename type="first">W</forename><surname>Kraaij</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Pohlmann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Hiemstra</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the eighth Text Retrieval Conference TREC-8</title>
				<meeting>the eighth Text Retrieval Conference TREC-8</meeting>
		<imprint>
			<publisher>NIST Special Publications</publisher>
			<date type="published" when="2000">2000</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Viewing stemming as recall enhancement</title>
		<author>
			<persName><forename type="first">W</forename><surname>Kraaij</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Pohlmann</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 19th ACM-SIGIR Conference on Research and Development in Information Retrieval (SIGIR96)</title>
				<editor>
			<persName><forename type="first">H</forename><forename type="middle">P</forename><surname>Frei</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">D</forename><surname>Harman</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Schäuble</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">R</forename><surname>Wilkinson</surname></persName>
		</editor>
		<meeting>the 19th ACM-SIGIR Conference on Research and Development in Information Retrieval (SIGIR96)</meeting>
		<imprint>
			<date type="published" when="1996">1996</date>
			<biblScope unit="page" from="40" to="48" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Parallel web corpora for CLIR?</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">Y</forename><surname>Nie</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of CLEF 2000</title>
				<meeting>CLEF 2000</meeting>
		<imprint>
			<date type="published" when="2000">2000</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Cross-language information retrieval based on parallel texts and automatic mining of parallel texts in the web</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">Y</forename><surname>Nie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Simard</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Isabelle</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Durand</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">ACM-SIGIR&apos;99</title>
				<imprint>
			<date type="published" when="1999">1999</date>
			<biblScope unit="page" from="74" to="81" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">The effect of syntactic phrase indexing on retrieval performance for Dutch texts</title>
		<author>
			<persName><forename type="first">R</forename><surname>Pohlmann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Kraaij</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of RIAO&apos;97</title>
				<editor>
			<persName><forename type="first">L</forename><surname>Devroye</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">C</forename><surname>Chrisment</surname></persName>
		</editor>
		<meeting>RIAO&apos;97</meeting>
		<imprint>
			<date type="published" when="1997">1997</date>
			<biblScope unit="page" from="176" to="187" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Relevance weighting of search terms</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">E</forename><surname>Robertson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><forename type="middle">Sparck</forename><surname>Jones</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of the American Society for Information Science</title>
		<imprint>
			<biblScope unit="volume">27</biblScope>
			<biblScope unit="page" from="129" to="146" />
			<date type="published" when="1976">1976</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
