<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Pitt at CLEF05: Data Fusion for Spoken Document Retrieval</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Daqing</forename><surname>He</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">School of Information Sciences</orgName>
								<orgName type="institution">University of Pittsburgh IS Building</orgName>
								<address>
									<addrLine>135 N. Bellefield Avenue</addrLine>
									<postCode>15260</postCode>
									<settlement>Pittsburgh</settlement>
									<region>PA</region>
									<country key="US">USA</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Jaewook</forename><surname>Ahn</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">School of Information Sciences</orgName>
								<orgName type="institution">University of Pittsburgh IS Building</orgName>
								<address>
									<addrLine>135 N. Bellefield Avenue</addrLine>
									<postCode>15260</postCode>
									<settlement>Pittsburgh</settlement>
									<region>PA</region>
									<country key="US">USA</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Pitt at CLEF05: Data Fusion for Spoken Document Retrieval</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">D1548B9EBDB07402AA5D47472668E5B5</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-25T00:39+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>H.3 [Information Storage and Retrieval]: H.3.3 Information Search and Retrieval Measurement</term>
					<term>Performance</term>
					<term>Experimentation Spoken Document Retrieval</term>
					<term>Data Fusion</term>
					<term>Evidence Combination</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>This paper describes an experimental investigation of data fusion techniques for spoken document retrieval. The effectiveness of retrievals solely based on the outputs from automatic speech recognition (ASR) is subject to the recognition errors introduced by the ASR process. This is especially true for retrievals on Malach test collection, whose ASR outputs have average word error rate (WER) of 35explored data fusion techniques for integrating the manually generated metadata information, which is provided for every Malach document, with the ASR outputs. We concentrated our effort on the post-search data fusion techniques, where multiple retrieval results using automatic generated outputs or human metadata were combined. Our initial studies indicated that a simple unweighted combination method (i.e., CombMNZ) that had demonstrated to be useful in written text retrieval environment only generated significant 38in retrieval effectiveness (measured by Mean Average Precision) for our task by comparing to a simple retrieval baseline where all manual metadata and ASR outputs are put together. This motivated us to explore a more elaborated weighted data fusion model, where the weights are associated with each retrieval results, and can be specified by the users in advance. We also explored multiple iteration of data fusion in our weighted fusion model, and obtained further improvement at 2nd iteration. In total, our best run on data fusion obtained 31baseline, and 4which is a significant difference.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>Spoken documents become more and more popular in people's information seeking activities along with the advance of information technologies, especially the storage and network communication technologies. However, comparing to the studies performed on text base documents, the achievements on retrieving spoken documents are still far less. Recent remarkable advances in the automatic recognition of spontaneous conversational speech makes it even urgent to study effective spoken document retrieval techniques. This is the reason that we participated CLEF Spoken Document Retrieval (SDR) track, and our goal is to leverage technologies developed for text retrieval into retrieving spoken documents.</p><p>Retrieving spoken documents from Malach test collection, a test collection developed by University of Maryland for retrieving spontaneous conversational speech <ref type="bibr" target="#b6">[8]</ref>, poses some interesting challenges. Before Malach collection, there have been several spoken document collections, whose documents are mostly news broadcast stories, help desk telephone calls, and political speeches <ref type="bibr" target="#b0">[1]</ref>. The documents in the Malach collection, however, are interviews of Holocaust survivors, which talk about their personal stories. Because of the genre, the topics, and the emotion involved, the average Word Error Rate (WER) in machine transcripts of the documents, an indicator of the quality of the ASR output, is around 35%. This imposes a great difficult for searching based on these ASR outputs.</p><p>Our major interests for this year's experiments, however, lie on the several forms of human generated metadata associated with the spoken documents. For each document, there are a list of person names mentioned in the documents, the human assigned thesaurus keywords and a brief summary in 2-3 sentences written by the human catalogers during their cataloging process.</p><p>We view ASR outputs and human generated metadata as two types of information that are complimentary of each other in retrieval process. On the one hand, ASR outputs provide full and detailed information about the content of the documents, which often could not be totally covered by human generated data. On the other hand, human generated metadata provide focused, humanprocessed, and high quality information that can be relied on for the accuracy of the retrieval. If we can develop a reliable retrieval method that can combine both information into the retrieval process in such way that their complimentary features can be fully explored, the achieved retrieval effectiveness would be greatly superior than that of any one of them. This is the goal of our studies in this year's CLEF-SDR experiments, and two derived research questions are:</p><p>1. how the manual metadata and ASR outputs in Malach collection can be integrated for improving the retrieval effectiveness of the final run?</p><p>2. what are the parameters that we can utilize to make the data fusion techniques more effective for our task?</p><p>In the rest of this report, we will firstly review some existing data fusion methods in Section 2; discuss in detail the experiment settings in Section 3; then talk about the fusion techniques we developed for the CLEF-SDR experiments in Section 4. Finally we will discuss some further studies in Section 5.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Data Fusion</head><p>In the literature, the techniques for combining multiple queries, document representations or retrieval results is called "data fusion" <ref type="bibr" target="#b5">[7]</ref>. It has been an active topic in text retrieval process, and people have developed many techniques for applying fusion techniques in various retrieval applications. Belkin et al. <ref type="bibr" target="#b1">[2]</ref> studied pre-search data-fusion approach by progressively combining Boolean query formulations. Lee <ref type="bibr" target="#b5">[7]</ref> provided an analysis of multiple post-search data fusion methods using TREC3 ad hoc retrieval data. Data fusion also has been applied in cross-language information retrieval <ref type="bibr" target="#b3">[4,</ref><ref type="bibr" target="#b2">3]</ref>, recommendation systems <ref type="bibr" target="#b7">[9]</ref>, and many other areas.</p><p>In post-search data fusion approaches, to properly merge retrieval results that are commonly ranklist of documents, the score associated with each document has to be normalized within that list. A often used normalization scheme (see Equation ( <ref type="formula">1</ref>)) utilizes the maximum and minimum scores of a ranklist (i.e., M axScore and M inScore) in the normalization process <ref type="bibr" target="#b5">[7]</ref>.</p><formula xml:id="formula_0">N ormalizedScore = U nnormalizedScore − M inScore M axScore − M inScore (1)</formula><p>Fox and Shaw <ref type="bibr" target="#b4">[5]</ref> developed several fusion methods for combining multiple evidence, and they named the methods as CombMIN, CombMAX, CombSUM, CombANZ, and CombMNZ (the definitions of them are shown in table <ref type="table" target="#tab_0">1</ref>). Lee <ref type="bibr" target="#b5">[7]</ref>   <ref type="bibr" target="#b6">[8]</ref>. The collection contains about 7800 segments from 300 interviews of Holocaust survivors. All the segments were constructed manually by catalogers. Each segment contains two automatic speech recognition outputs from the ASR system developed by IBM in 2003 and 2004 respectively (see Figure <ref type="figure" target="#fig_0">1</ref> for an example of the output from ASR 2004 system). The WER of the two outputs are about 40% and 35% respectively. In addition, there are automatically generated thesaurus terms from a system developed at University of Maryland. Each segment also contains a set of human generated data, including person names mentioned in the segment, average 5 thesaurus labels and 3-sentence summaries (see Figure <ref type="figure">2</ref> for an example).</p><p>There are total 63 search topics, 38 of which were available for training, and 25 were held as the testing topics. Each topic is designed in TREC style, which has a title, a description and a narrative (see Figure <ref type="figure" target="#fig_1">3</ref>). The topics are available in English, Spanish, Czech, German and French, however, we only used English topics for our studies.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2">Indri search engine</head><p>Our search engine was Indri 1.0, which was a collaboration effort between the University of Massachusetts and Carnegie Mellon University [6]. Its retrieval model is a combination of language model and inference network. We chose it not only because of its state-of-art retrieval effectiveness, but also because it is an open source system, on which we can easily integrated our modifications. Its powerful query syntax is another attraction to us, since we want to specify which index fields should a retrieval be based on for our studies of manual metadata only or automatic data only searches.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3">Measures</head><p>To study the retrieval results in as wide scenarios as possible, instead of choosing on single measure, we employed a set of evaluation measures, each of which tells us some aspect of the retrieval effectiveness of the search results:</p><p>• mean average precision (MAP), the measure aim at giving an emphasis view of precision in a ranklist. Since the ranks of the relevant documents are considered in the measure, this measure gives a reasonable overview of the quality of the ranklist for a given retrieval topic.</p><p>&lt;ASRTEXT2004A&gt; oh i'll you know are yeah yeah yeah yeah yeah yeah yeah the very why don't we start with you saying anything in your about grandparents great grandparents well as a small child i remember only one of my grandfathers and his wife his second wife he was selling flour and the type of business it was he didn't even have a store he just a few sacks of different flour and the entrance of an apartment building and people would pass by everyday and buy a chela but two killers of flour we have to remember related times were there was no already baked bread so people had to baked her own bread all the time for some strange reason i do remember fresh rolls where everyone would buy every day but not the bread so that was the business that's how he made a living where was this was the name of the town it wasn't shammay dish he ours is we be and why i as i know in southern poland and alisa are close to her patient mountains it was rather mid sized town and uhhuh i was and the only child and the family i had a governess who was with me all their long from the time i got up until i went to sleep she washed me practice piano she took me to ballet lessons she took me skiing and skating wherever there was else that i was doing being non reach higher out i needed other children to players and the governors were always follow me and stay with me while ours twang that i was a rotten spoiled care from work to do family the youngest and the large large family and everyone was door in the army &lt;/ASRTEXT2004A&gt; • Precision at 10 (P10) is a useful measure to examine how much relevant documents are there in the first result screen, which is often the only results viewed by a user.</p><p>• R-Precision (R-PREC) emphases on precision but also avoids the artificial cut-off effect imposed by pre-defined cut-off point, like in P10. The "R" varies according to the number of relevant documents of a given topic.</p><p>• Average Recall at top 1000 returned documents. It tells the quality of the ranklist from the point of recall.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.4">Baselines</head><p>We established three baselines for evaluating our methods. The first two represent the scenario that no data fusion is performed. We selected a run on ASRTEXT 2004 as the baseline for searching on ASR output, since ASRTEXT 2004 is the better one among the two ASR outputs. This baseline is referred as "asr04", and is treated as the lower baseline. Ideally, we should use  the search on manual transcripts as the upper baseline. Since Malach collection does not provide manual transcripts, we used all manually generated data in the segments as the surrogate. It acts as a proximate upper bound baseline, and is referred as "manual-only" baseline. We did not apply blind relevance feedback (BRF) over "manual-only" run since our explorations using Indri's BRF function over "manual-only" baseline generated inferior results. The third baseline is a search on all manual and ASR outputs being put together as if they are different parts of the same document. This represents the simplest data fusion method, and is referred "simple-fusion" baseline.</p><p>The search effectiveness of the three baselines are presented in The first data fusion method studied in the CLEF-SDR experiments was our implementation of CombMNZ method since Lee demonstrated its superior over the other three methods <ref type="bibr" target="#b5">[7]</ref>. The run based on CombMNZ merged results from three retrieval runs, the "manual-only" baseline, the "asr04" baseline, and a retrieval run on the automatic assigned thesaurus keywords called "AU-TOKEYWORD2004A1" (we call this run "autokw1"). Table <ref type="table" target="#tab_5">3</ref> shows the results of "CombMNZ" run and that of the three runs that it was based on. Comparing to the lower "asr04" baseline, this combined run has achieved significant improvement by the measures of MAP, P10, and especially Avg-Recall (P 0.05 in paired T-tests). However, it generates significant decrease at MAP, R-PREC, and P10 when comparing to the two higher baselines, "manual-only" baseline and "simple-fusion" baseline (P 0.05 in paired T-test). The only improvement it achieved over the two higher baselines is measured by Avg-Recall. This means that "CombMNZ" run does return more relevant documents comparing to the two high baselines, but it ranks them badly.</p><p>A close exam of the retrieval runs in Table <ref type="table" target="#tab_5">3</ref> shows that the retrieval effectiveness of "manualonly" run is greatly higher than that of the two automatic runs. For example, the MAP of "manual-only" increases about 200% over "asr04", the better one of the two automatic runs. Therefore, it makes no sense to assume that their contribution to the final fused ranklist is the same, which is the assumption in CombMNZ model. We need a data fusion model that considers the retrieval difference.  </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2">Weighted CombMNZ</head><p>The failure of CombMNZ on our data fusion task motivated us to explore a weighted scheme for data fusion based on CombMNZ. A natural place to insert a weight in CombMNZ is to assign a weight of belief for each retrieval run as it is possible to obtain such evidence or belief. In our weighted CombMNZ model (called WCombMNZ model), such belief is used in calculation of the final combined score for a document (see Equation <ref type="formula">2</ref>).</p><formula xml:id="formula_1">W CombM N Z i = n j=1 (w j × N ormalizedScore i,j ) × n (2)</formula><p>where w j is a predefined weight associated with a search result to be combined, n is the number of nonzero scores of document i, and the N ormalizedScore i,j is calculated using Equation <ref type="formula">1</ref>. Various methods can be used to obtain the weight w j for a given ranklist j. In this year's experiment, we firstly used the retrieval effectiveness of those pre-combined runs on the 38 training topics as the weights (the details of the pre-fused runs on the training topics are in Table <ref type="table" target="#tab_6">4</ref>). Therefore, we have four different "WCombMNZ" runs (see Table <ref type="table" target="#tab_7">5</ref>), and their retrieval effectiveness evaluated by the four measures are in Table <ref type="table" target="#tab_8">6</ref>.  All four WCombMNZ runs are significant higher than the non-weighted "CombMNZ" run (paired T-test with P &lt; 0.05) when looking at MAP, R-PREC, and P10 as the measures. However, they are still significant lower than the "manual-only" run using the same measures.</p><p>Since the weights of "WCombMNZ-1" generated the best MAP, R-PREC and P10 results, we used those weights to help us explored the effect of different combinations of the weights. As the difference of the retrieval effectiveness between "manual-only" run and the two automatic runs is significant higher than that between the two automatic runs, we first explored the change of ratio between the weight of "manual-only" run and that of the "asr-04" and "autowa1" runs. The ratio we tested were 2:1 (that is the weight for "manual-only" run is 2, and the weights for the two automatic runs were both assigned to be 1 in WCombMNZ model), 5:1, 10:1, 15:1, and up to 1000:1 (the results are presented in Table <ref type="table" target="#tab_9">7</ref>). We then change the weight ratio between the "asr04" and that of "autowa1" to 2:1, which is closer to the weight ratio in "WCombMNZ-1", and varied the weight ratio of the three runs from 4:2:1, 6:2:1, and up to 50:2:1. As shown in Table <ref type="table" target="#tab_9">7</ref>, the ratio between the weight of the "manual-only" and that of two automatic runs is the dominate factor in affecting the retrieval performance, and once the ratio between the manual run and the automatic runs is larger than 10, there is not much difference in the retrieval effectiveness evaluated by all measures. However, still none of the fused runs achieves better MAP, R-PREC, and P10 than "manual-only", although they are much closer to the performance of the "manualonly" than the two automatic runs, and at the same time, many of they have achieved significant better Avg-Recall than the "manual-only" run.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.3">Multiple Iteration of Data Fusion</head><p>One exploration within the data fusion framework is 'does multiple iterations of data fusion make sense?" To answer this, we conducted several experiments in WCombMNZ model. Total five retrieval runs were used in the second iteration of data fusion. We kept the "manual-only" run since it is the best run so far, and we used the four runs listed in Table <ref type="table" target="#tab_8">6</ref> "WCombMNZ-1" to "WCombMNZ-4". We used the similar scheme to vary the weight ratios among the runs, and we also set all weights to 1 to make WCombMNZ model fall back to CombMNZ so that we can study CombMNZ too. The ratio "2-1" in Table <ref type="table" target="#tab_10">8</ref> means that the weight for "manual-only" is 2, and that for the other four runs is 1.</p><p>As shown in Table <ref type="table" target="#tab_10">8</ref>, the "manual-only", in our current retrieval setting, still deserves more weights than the other runs, and the best retrieval results are achieved with the ratio around 10:1. Statistical tests (paired T-test) between the results of the 2nd round fusion runs and that of the "manual-only" run demonstrate that all the 2-iteration fusion data generated significant improvement on average recall, but only the runs with ratio above 10:1 generated significant  We then generated various 3rd round fusion runs using the similar scheme, which include the "manual-only" run, and the four 2nd round runs with the ratio 5:1, 10:1, 15:1 and 20:1. The results are shown in Table <ref type="table" target="#tab_10">8</ref>. None of the 3rd round runs could generate statistical significant improvement over the 2nd round runs. It seems that fusions with iteration more than 2 does not justify the extra costs involved comparing to the 2nd round fusion runs.</p><p>Our multiple iteration experiments tell us that it is usually difficult to obtain a better fusion results over the best pre-fusion run when the retrieval effectiveness of the pre-fusion runs are greatly different to each other. The fusion experiment on "manual-only" and the other automatic runs is an example of such fusion. However, significant improvement over the best pre-fusion run could be achieved via multiple iterations of fusion. For example, we achieved significant improvement over the best pre-fusion run in two iterations. Of course, we need further experiments to test the general effectiveness of the multiple iteration fusion.</p><p>In this paper, we have described an experimental investigation of data fusion techniques for spoken document retrieval. Because of the various characteristics of the documents in the Malach test collection, retrieval solely based on the outputs from automatic speech recognition (ASR) is well below retrievals on manual generated data. To overcome the problem, we have explored data fusion techniques for integrating the manually generated metadata information with the ASR outputs. We concentrated on the post-search fusion approach, and explored weighted CombMNZ model with different weight ratios and multiple iterations. Our initial results indicate that a simple unweighted combination method that has demonstrated to be useful in written retrieval environment <ref type="bibr" target="#b5">[7]</ref> only generated significant 38% relative decrease in retrieval effectiveness (Mean Average Precision) for our task by comparing to a simple retrieval baseline where all manual metadata and ASR outputs are put together. Only with the more elaborated weighted combination scheme did we obtained 31% significant relative improvements over the simple fusion baseline, and 4% relative improvement over the manual-only baseline, which is a significant difference.</p><p>Our future work include further experiments on the general effectiveness of the multiple iteration fusion, and another future work is to explore the usage of WCombMNZ in other retrieval tasks, where multiple retrieval results can be obtained from one retrieval engine, or even different engines. The third further study we want to work on is to answer the question what is the minimum human generated data to ASR output if the goal is to combine the human generated data with the ASR output to achieve a comparable retrieval effectiveness to a retrieval on manual transcripts.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: An example of ASR output (2004 version) and automatically generated thesaurus terms (autowa1 version)</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 3 :</head><label>3</label><figDesc>Figure 3: An example of the search topic in Malach Collection</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>WCombMNZ- 1</head><label>1</label><figDesc>use the MAP values of the three runs on training topics as the weights WCombMNZ-2 use the R-PREC values of the three runs on training topics as the weights WCombMNZ-3 use the P10 values of the three runs on training topics as the weights WCombMNZ-4 use the Avg-Recall values of the three runs on training topics as the weights</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1 :</head><label>1</label><figDesc>studied these methods, and established that CombMNZ is the best among the four in retrieving TREC ad hoc data. Combining functions proposed by Fox and Shaw</figDesc><table><row><cell>CombMIN</cell><cell>minimum of all scores of a document</cell></row><row><cell cols="2">CombMAX maximum of all scores of a document</cell></row><row><cell cols="2">CombSUM summation of all scores of a document</cell></row><row><cell cols="2">CombANZ CombSUM ÷ number of nonzero scores of a document</cell></row><row><cell cols="2">CombMNZ CombSUM × number of nonzero scores of a document</cell></row><row><cell cols="2">3 Experiment Settings</cell></row><row><cell cols="2">3.1 Malach Test Collection</cell></row><row><cell cols="2">Malach Test Collection was developed by University of Maryland as part of their effort in Malach</cell></row><row><cell>project</cell><cell></cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 2</head><label>2</label><figDesc></figDesc><table><row><cell>runs</cell><cell>MAP</cell><cell>R-PREC</cell><cell>P10</cell><cell>Avg-Recall</cell></row><row><cell cols="2">manual-only 0.2312</cell><cell>0.2836</cell><cell>0.4603</cell><cell>0.5813</cell></row><row><cell>asr04</cell><cell>0.0693</cell><cell>0.1139</cell><cell>0.2111</cell><cell>0.3595</cell></row><row><cell cols="2">simple-fusion 0.1842</cell><cell>0.1985</cell><cell>0.3635</cell><cell>0.5847</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 2 :</head><label>2</label><figDesc>Retrieval effectiveness of the three baselines</figDesc><table><row><cell>4 Experiments and Results Analysis</cell></row></table><note>4.1 Data Fusion with CombMNZ</note></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_5"><head>Table 3 :</head><label>3</label><figDesc>Retrieval effectiveness of CmbMNZ run</figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_6"><head>Table 4 :</head><label>4</label><figDesc>Retrieval effectiveness of individual runs on the 38 training topics</figDesc><table><row><cell>runs</cell><cell>MAP</cell><cell>R-PREC</cell><cell>P10</cell><cell>Avg-Recall</cell></row><row><cell cols="2">manual-only 0.1494</cell><cell>0.1823</cell><cell>0.3237</cell><cell>0.4221</cell></row><row><cell>asr04</cell><cell>0.0525</cell><cell>0.0754</cell><cell>0.1447</cell><cell>0.2788</cell></row><row><cell>autowa1</cell><cell>0.0239</cell><cell>0.0460</cell><cell>0.0816</cell><cell>0.2832</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_7"><head>Table 5 :</head><label>5</label><figDesc>Combining functions proposed by Fox and Shaw</figDesc><table><row><cell>runs</cell><cell>MAP</cell><cell>R-PREC</cell><cell>P10</cell><cell>Avg-Recall</cell></row><row><cell>manual-only</cell><cell>0.2312</cell><cell>0.2836</cell><cell>0.4603</cell><cell>0.5813</cell></row><row><cell>CombMNZ</cell><cell>0.1127</cell><cell>0.1173</cell><cell>0.3079</cell><cell>0.6182</cell></row><row><cell cols="2">WCombMNZ-1 0.2137</cell><cell>0.2589</cell><cell>0.4460</cell><cell>0.6008</cell></row><row><cell cols="2">WCombMNZ-2 0.1987</cell><cell>0.2431</cell><cell>0.4206</cell><cell>0.6254</cell></row><row><cell cols="2">WCombMNZ-3 0.1967</cell><cell>0.2416</cell><cell>0.4190</cell><cell>0.6253</cell></row><row><cell cols="2">WCombMNZ-4 0.1783</cell><cell>0.2188</cell><cell>0.3778</cell><cell>0.6215</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_8"><head>Table 6 :</head><label>6</label><figDesc>Retrieval effectiveness of The first 4 WCmbMNZ runs on total 63 topics</figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_9"><head>Table 7 :</head><label>7</label><figDesc>Exploring the weight ratios in WCmbMNZ model improvement on P10, and only runs with ratio 10:1 and 15:1 generated significant improvement on MAP. No significant improvement can be achieved on R-PREC.</figDesc><table><row><cell>runs</cell><cell>MAP</cell><cell>R-PREC</cell><cell>P10</cell><cell>Avg-Recall</cell></row><row><cell>ratio 2-1</cell><cell>0.1884</cell><cell>0.2315</cell><cell>0.4032</cell><cell>0.6236</cell></row><row><cell>ratio 5-1</cell><cell>0.2088</cell><cell>0.2590</cell><cell>0.4254</cell><cell>0.6259</cell></row><row><cell>ratio 10-1</cell><cell>0.2133</cell><cell>0.2598</cell><cell>0.4302</cell><cell>0.6240</cell></row><row><cell>ratio 15-1</cell><cell>0.2132</cell><cell>0.2590</cell><cell>0.4381</cell><cell>0.6211</cell></row><row><cell>ratio 20-1</cell><cell>0.2140</cell><cell>0.2581</cell><cell>0.4413</cell><cell>0.6202</cell></row><row><cell>ratio 25-1</cell><cell>0.2131</cell><cell>0.2581</cell><cell>0.4444</cell><cell>0.6129</cell></row><row><cell>ratio 50-1</cell><cell>0.2141</cell><cell>0.2577</cell><cell>0.4429</cell><cell>0.6133</cell></row><row><cell>ratio 100-1</cell><cell>0.2140</cell><cell>0.2583</cell><cell>0.4460</cell><cell>0.6087</cell></row><row><cell>ratio 1000-1</cell><cell>0.2132</cell><cell>0.2593</cell><cell>0.4460</cell><cell>0.5995</cell></row><row><cell>ratio 4-2-1</cell><cell>0.1937</cell><cell>0.2354</cell><cell>0.3735</cell><cell>0.6246</cell></row><row><cell>ratio 6-2-1</cell><cell>0.2047</cell><cell>0.2523</cell><cell>0.4190</cell><cell>0.6253</cell></row><row><cell>ratio 10-2-1</cell><cell>0.2110</cell><cell>0.2591</cell><cell>0.4238</cell><cell>0.6236</cell></row><row><cell>ratio 20-2-1</cell><cell>0.2138</cell><cell>0.2599</cell><cell>0.4333</cell><cell>0.6208</cell></row><row><cell>ratio 30-2-1</cell><cell>0.2144</cell><cell>0.2591</cell><cell>0.4365</cell><cell>0.6207</cell></row><row><cell>ratio 50-2-1</cell><cell>0.2141</cell><cell>0.2574</cell><cell>0.4444</cell><cell>0.6172</cell></row><row><cell cols="2">ratio 100-2-1 0.2140</cell><cell>0.2575</cell><cell>0.4444</cell><cell>0.6089</cell></row><row><cell>runs</cell><cell>MAP</cell><cell>R-PREC</cell><cell>P10</cell><cell>Avg-Recall</cell></row><row><cell>manual-only</cell><cell>0.2312</cell><cell>0.2836</cell><cell>0.4603</cell><cell>0.5813</cell></row><row><cell>2nd-ratio 1-1</cell><cell>0.2119</cell><cell>0.2670</cell><cell>0.4413</cell><cell>0.6241</cell></row><row><cell>2nd-ratio 2-1</cell><cell>0.2295</cell><cell>0.2720</cell><cell>0.4540</cell><cell>0.6228</cell></row><row><cell>2nd-ratio 5-1</cell><cell>0.2397</cell><cell>0.2806</cell><cell>0.4778</cell><cell>0.6200</cell></row><row><cell cols="2">2nd-ratio 10-1 0.2409</cell><cell>0.2860</cell><cell>0.4937</cell><cell>0.6188</cell></row><row><cell cols="2">2nd-ratio 15-1 0.2400</cell><cell>0.2853</cell><cell>0.4968</cell><cell>0.6157</cell></row><row><cell>2nd-ratio 20-1</cell><cell>0.2388</cell><cell>0.2856</cell><cell>0.4810</cell><cell>0.6142</cell></row><row><cell>3rd-ratio 1-1</cell><cell>0.2403</cell><cell>0.2876</cell><cell>0.5016</cell><cell>0.6143</cell></row><row><cell>3rd-ratio 2-1</cell><cell>0.2393</cell><cell>0.2865</cell><cell>0.4857</cell><cell>0.6142</cell></row><row><cell>3rd-ratio 1-2</cell><cell>0.2409</cell><cell>0.2866</cell><cell>0.4984</cell><cell>0.6157</cell></row><row><cell>3rd-ratio 1-5</cell><cell>0.2407</cell><cell>0.2867</cell><cell>0.4937</cell><cell>0.6159</cell></row><row><cell>3rd-ratio 1-10</cell><cell>0.2408</cell><cell>0.2869</cell><cell>0.4921</cell><cell>0.6168</cell></row><row><cell>3rd-ratio 1-15</cell><cell>0.2408</cell><cell>0.2864</cell><cell>0.4921</cell><cell>0.6168</cell></row><row><cell>3rd-ratio 1-30</cell><cell>0.2404</cell><cell>0.2869</cell><cell>0.4921</cell><cell>0.6171</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_10"><head>Table 8 :</head><label>8</label><figDesc>Exploring multiple iterations in data fusion</figDesc><table /></figure>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgment</head><p>The authors would like to thank Doug Oard, Gareth Jones and Ryen White for their tireless efforts to coordinate CLEF-SDR.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<ptr target="http://www.dcs.shef.ac.uk/spandh/projects/swag/" />
		<title level="m">Eu-su working group on spoken-word audio collections</title>
				<imprint>
			<date type="published" when="2003">2003</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">The effect of multiple query representations on information retrieval system performance</title>
		<author>
			<persName><forename type="first">N</forename><forename type="middle">J</forename><surname>Belikn</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Cool</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><forename type="middle">B</forename><surname>Croft</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">P</forename><surname>Callan</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceeding of ACM-SIGIR&apos;93</title>
				<meeting>eeding of ACM-SIGIR&apos;93</meeting>
		<imprint>
			<date type="published" when="1993">1993</date>
			<biblScope unit="page" from="339" to="346" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Cross-language retrieval experiments at clef 2002</title>
		<author>
			<persName><forename type="first">Aitao</forename><surname>Chen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of CLEF 2002</title>
				<meeting>CLEF 2002</meeting>
		<imprint>
			<date type="published" when="2002">2002</date>
			<biblScope unit="page" from="28" to="48" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">CLIR Experiments at Maryland for TREC 2002: Evidence Combination for Arabic-English Retrieval</title>
		<author>
			<persName><forename type="first">Kareem</forename><surname>Darwish</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Douglas</forename><forename type="middle">W</forename><surname>Oard</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of TREC 2002</title>
				<meeting>TREC 2002</meeting>
		<imprint>
			<date type="published" when="2002">2002</date>
			<biblScope unit="page" from="703" to="710" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Combination of multiple searches</title>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">A</forename><surname>Fox</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">A</forename><surname>Shaw</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proeddings of the 2nd Text REtrieval Conference (TREC-2)</title>
				<imprint>
			<date type="published" when="1994">1994</date>
			<biblScope unit="page" from="243" to="252" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Analyses of multiple evidence combination</title>
		<author>
			<persName><forename type="first">Joon</forename><surname>Ho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Lee</forename></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceeding of SIGIR&apos;97</title>
				<meeting>eeding of SIGIR&apos;97</meeting>
		<imprint>
			<date type="published" when="1997">1997</date>
			<biblScope unit="page" from="267" to="276" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Building an information retrieval test collection for spontaneous conversational speech</title>
		<author>
			<persName><forename type="first">Douglas</forename><forename type="middle">W</forename><surname>Oard</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Dagobert</forename><surname>Soergel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">David</forename><surname>Doermann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Xiaoli</forename><surname>Huang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">Craig</forename><surname>Murray</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jianqiang</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Bhuvana</forename><surname>Ramabhadran</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Martin</forename><surname>Franz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Samuel</forename><surname>Gustman</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of SIGIR&apos;94</title>
				<meeting>SIGIR&apos;94</meeting>
		<imprint>
			<date type="published" when="2004">2004</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Combination of evidence in recommendation systems characterized by distance functions</title>
		<author>
			<persName><forename type="first">M</forename><surname>Luis</surname></persName>
		</author>
		<author>
			<persName><surname>Rocha</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2002 World Congress on Computational Intelligence, FUZZ-IEEE&apos;02</title>
				<meeting>the 2002 World Congress on Computational Intelligence, FUZZ-IEEE&apos;02</meeting>
		<imprint>
			<publisher>IEEE Press</publisher>
			<date type="published" when="2002">2002</date>
			<biblScope unit="page" from="203" to="208" />
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
