<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Dublin City University at CLEF 2004: Experiments with the ImageCLEF St Andrew&apos;s Collection</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Gareth</forename><forename type="middle">J F</forename><surname>Jones</surname></persName>
							<email>gjones@computing.dcu.ie</email>
							<affiliation key="aff0">
								<orgName type="department">School of Computing</orgName>
								<orgName type="institution">Dublin City University</orgName>
								<address>
									<settlement>Dublin 9</settlement>
									<country key="IE">Ireland</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Declan</forename><surname>Groves</surname></persName>
							<email>dgroves@computing.dcu.ie</email>
							<affiliation key="aff0">
								<orgName type="department">School of Computing</orgName>
								<orgName type="institution">Dublin City University</orgName>
								<address>
									<settlement>Dublin 9</settlement>
									<country key="IE">Ireland</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Anna</forename><surname>Khasin</surname></persName>
							<email>akhasin@computing.dcu.ie</email>
							<affiliation key="aff0">
								<orgName type="department">School of Computing</orgName>
								<orgName type="institution">Dublin City University</orgName>
								<address>
									<settlement>Dublin 9</settlement>
									<country key="IE">Ireland</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Adenike</forename><surname>Lam-Adesina</surname></persName>
							<email>adenike@computing.dcu.ie</email>
							<affiliation key="aff0">
								<orgName type="department">School of Computing</orgName>
								<orgName type="institution">Dublin City University</orgName>
								<address>
									<settlement>Dublin 9</settlement>
									<country key="IE">Ireland</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Bart</forename><forename type="middle">Mellebeek Andy</forename><surname>Way</surname></persName>
							<email>away@computing.dcu.ie</email>
							<affiliation key="aff0">
								<orgName type="department">School of Computing</orgName>
								<orgName type="institution">Dublin City University</orgName>
								<address>
									<settlement>Dublin 9</settlement>
									<country key="IE">Ireland</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Dublin City University at CLEF 2004: Experiments with the ImageCLEF St Andrew&apos;s Collection</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">D469186A24783590EB9F06BB14D6B3D4</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T17:51+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>For the CLEF 2004 ImageCLEF St Andrew's Collection task the Dublin City University group carried out three sets of experiments. We carried out standard cross-language information retrieval (CLIR) runs using topic translation using machine translation (MT), combination of this run with image matching results from the VIPER system, and a novel document rescoring approach based on automatic MT evaluation metrics. Our standard CLIR approaches works well in comparison on this task. Encouragingly combination with image matching lists can produce small positive changes in the overall retrieval output. However, rescoring using the MT evaluation metrics in their current form significantly reduces retrieval effectiveness.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>Dublin City University's participation in the CLEF 2004 ImageClef St Andrew's task comprised three sets of experiments for Dutch, French, German, Italian and Spanish topic languages. First, we explored the application of our existing CLIR system used in previous CLEF workshops <ref type="bibr" target="#b0">[1]</ref> with topic translation using three web-based translation resources. Second, combined this with the image matching results resulted provided by the track organisers' generated using the VIPER system. Finally, we explored a novel approach to rescoring the potentially relevant documents retrieved using our standard system based on automatic machine translation (MT) evaluation metrics.</p><p>The paper briefly outlines the details of our retrieval system before giving results for the first two sets of experiments. A separate section describes the MT evaluation metrics and gives results.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Standard CLIR Methodology 2.1 Retrieval System</head><p>The basis of our experimental retrieval system is the City University research distribution version of the Okapi system, as used in our previous CLEF participation <ref type="bibr" target="#b0">[1]</ref>. The documents and search topics were processed to remove stopwords from a list of about 260 words; suffix stripped using the Okapi implementation of Porter stemming and terms were indexed using a small set of synonyms.</p><p>Terms are weighted using the standard BM25 weighting scheme and all runs use our summarybased pseudo relevance feedback (PRF) method <ref type="bibr" target="#b1">[2]</ref>. The summary generation method combines Luhn's keyword cluster method, a title terms frequency method, a location/header method and a query-bias method from to form an overall significance score for each sentence. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2">Experimental Results</head><p>For the all the experiments report here the Okapi parameters were set using the provided training topics as follows; K1 = 1.0 and b = 0.5 for baseline runs and K1 = 1.5 and b = 0.6 for PRF runs. The 20 top ranked PRF expansion terms were selected from the summaries of the top 5 ranked documents. The original topic terms were upweighted by a factor of 3.5 relative to terms introduced by PRF. There are a total of 829 relevant images available in the collection. Topics were translated into English, the document language, using the following web-based MT systems: Systran (http://www.systransoft.com/), SDL (http://www.freetranslation.com/) and InterTrans (http://www.intertrans.com/). Results are show for CLIR using each separate translation and a term union merged translation.</p><p>Baseline Runs Table <ref type="table" target="#tab_0">1</ref> shows baseline retrieval runs for Systran without application of PRF. Results in all languages appear reasonable with little apparent correlation between precision and recall figures.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Feedback Runs</head><p>The text annotations of the images are typically very short, typically comprising only a few sentences. In developing our system for the PRF compared our summary-based approach developed for use with newspaper archives with a standard PRF approach selecting terms from complete documents. We were a little surprised to find that selecting terms from summaries of even these short documents worked better on the development topics than the whole document approach.</p><p>Table <ref type="table" target="#tab_1">2</ref> shows feedback results for each topic language with the three MT systems and the merged translated topics. Separate results are shown for the original and later released revised Spanish topics. Comparing all these runs we can see that for Systran, PRF on on average produces an improvement in average precision for each language pair, although there is no clear trend for relevant document recall. Comparing between the alternative topic translations it can be seen that different systems on average produce the best average precision for different language pairs, although in general InterTrans is the least effective. Results for the merged topics are rather mixed. It was hoped that the increased term coverage would improve recall and aid precision; this does happen in some cases, but in others it reduces effectiveness. Further investigation is needed into specific success and failures to see if any general conclusions can be made.</p><p>Text and Image Combination Runs Table <ref type="table" target="#tab_2">3</ref> shows results for a simple sum combination of the matching score for the PRF runs shown in Table <ref type="table" target="#tab_1">2</ref> and the provided VIPER runs. These results are only slightly different from the PRF runs. However, some potentially important positives can be taken from this. First, in image retrieval it is often found that adding image matching information cannot improve over text caption only retrieval. For our experiments in some cases the image matching score does help, albeit only marginally. Second, the VIPER system was not adjusted for the St Andrew's collection task, suggesting that a better image matching run should be possible with some task specific training of the image matching process.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Machine Translation Quality Metric Runs</head><p>In recent years, several automatic MT evaluation methods have been proposed as a supplement to, or, in certain cases, a replacement for costly human MT evaluations <ref type="bibr" target="#b2">[3]</ref>[4][5] <ref type="bibr" target="#b5">[6]</ref>. These automatic evaluation methods rely on the idea that the quality of an MT can be measured by its similarity to a professional human translation. With each of the currently available automatic evaluation methods, this similarity is measured using a word-error metric between the sentences in the MTproduced text and the sentences in one or more human reference translations. The success of automatic MT evaluation depends largely on the amount of available comparable material and on the number of human reference translations, with more reference translations resulting in a more accurate measure of system performance.</p><p>In order to be able to use these metrics to calculate the similarity between a user query and a topic document in IR, we regard the original topic document and the MT-translated user query as translations of an unknown source text, as is shown in Figure <ref type="figure" target="#fig_0">1</ref>.</p><p>The same three sets of topic translation were used as in the previous experiments. The topic translations and documents were pre-processed to remove stopwords, capitalisation and punctuation.</p><p>If we think of the query translations as human reference translations, it is possible to measure the accuracy of the would-be 'machine translations' (the documents) using automatic MT evaluation metrics. The best 'machine translation' is the translation with the lowest word-error score with regard to the reference translations. The goal of our experiment was to find out to what extent the best 'machine translation' corresponded with a relevant document. Experiments with development topics showed that best results were obtained with a combination of 2 existing MT evaluation methods (NIST and GTM) and an adaptation of the BLEU evaluation metric.</p><p>BLEU ranks different MT output texts based a combination of an N-gram similarity score and a sentence brevity penalty with respect to a corpus of human reference translations. The BLEU evaluation script was adapted in two ways. First, we eliminated the sentence brevity penalty. The original BLUE metric penalizes short sentences to avoid the possibility that very short segments such as 'the' would receive a maximum score when compared to any sentence containing 'the'. This penalty is clearly not relevant for the retrieval task at hand. A second modification to the script consisted in allowing a non-zero BLEU score, regardless of the fact that for one or more of the N-gram categories (unigram to 4-gram) no positive matches were found between the MT output and human reference translations.</p><p>NIST differs from BLEU with respect to both the co-occurrence score and the sentence brevity penalty. NIST alters the co-occurrence score in favour of lower order N-grams (i.e. low trigrams or quadrigram matches play less a role in the overall score) and more informative N-grams (i.e. N-grams that occur less frequently receive a higher weight). The sentence brevity penalty used by NIST is less severe than the one used by BLEU for sentences with small variations with respect to the reference translation.</p><p>GTM allows the calculation of standard precision and recall scores for automatically produced translations. It also calculates an f-measure score, which combines both the precision and recall scores for a given translation. It is this f-measure score, along with the NIST and adapted BLEU scores, that we used in our automatic ranking of the documents.</p><p>During our experiments we ranked the translated queries against the top 1000 documents retrieved for each topic using the PRF approach described in the previous section. We used a summation of the NIST, f-measure and adapted BLEU scores. We ran two sets of experiments. In the first set of experiments we evaluated the retrieved document list against only one reference translation, as produced by one of the three online MT systems, giving us three resulting ranking lists of documents for each topic. In a second set of experiments we merged the translated queries, using the three different translations of the topic as three different reference translations.</p><p>Table <ref type="table" target="#tab_3">4</ref> shows results of document rescoring using MT evaluation metrics. Comparing these results to those using standard PRF methods in the earlier tables, it can be seen that the MT evaluation metrics are not effective for IR scoring in their present form. The main goal of our experiments was not to substantially improve the best available Image Retrieval methods, but to investigate the novel idea for IR of treating topic documents and translated user queries as comparable translations of an unknown source text. Clearly based on the results shown here we need to explore further whether this approach can be adapted successfully for IR applications.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Conclusions and Further Work</head><p>Our experiments for ImageCLEF have demonstrated that our standard CLIR method works effectively for the short text documents in the St Andrew's collection, and further that there is potential for improvement in retrieval effectiveness from the use of image matching in CL image retrieval.</p><p>Our experiments using MT evaluation metrics for scoring CLIR have so far not been successful, but we will be analysing our results to better understand the results and to seek alternative means of applying this approach.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Document scoring based on MT Evaluation metrics.</figDesc><graphic coords="4,131.42,108.86,340.16,158.49" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1 :</head><label>1</label><figDesc>Baseline retrieval runs using Systran topic translation.</figDesc><table><row><cell></cell><cell cols="6">Dutch French German Italian Spanish</cell></row><row><cell cols="2">Prec. 10 docs 0.424</cell><cell>0.488</cell><cell>0.536</cell><cell></cell><cell>0.396</cell><cell>0.416</cell></row><row><cell>Av Precision</cell><cell>0.384</cell><cell>0.427</cell><cell>0.464</cell><cell></cell><cell>0.402</cell><cell>0.383</cell></row><row><cell>Rel. Ret.</cell><cell>698</cell><cell>631</cell><cell>695</cell><cell></cell><cell>606</cell><cell>654</cell></row><row><cell></cell><cell></cell><cell></cell><cell>SDL</cell><cell>INT</cell><cell>ST</cell><cell>MG</cell></row><row><cell>Dutch</cell><cell cols="6">Prec. 10 docs 0.472 0.276 0.500 0.472</cell></row><row><cell></cell><cell cols="6">Av Precision 0.398 0.273 0.432 0.421</cell></row><row><cell></cell><cell cols="2">Rel. Ret.</cell><cell>683</cell><cell>637</cell><cell>709</cell><cell>791</cell></row><row><cell>French</cell><cell cols="6">Prec. 10 docs 0.472 0.532 0.496 0.432</cell></row><row><cell></cell><cell cols="6">Av Precision 0.409 0.466 0.431 0.399</cell></row><row><cell></cell><cell cols="2">Rel. Ret.</cell><cell>666</cell><cell>707</cell><cell>658</cell><cell>695</cell></row><row><cell>German</cell><cell cols="6">Prec. 10 docs 0.592 0.528 0.540 0.632</cell></row><row><cell></cell><cell cols="6">Av Precision 0.501 0.468 0.474 0.531</cell></row><row><cell></cell><cell cols="2">Rel. Ret.</cell><cell>763</cell><cell>804</cell><cell>691</cell><cell>804</cell></row><row><cell>Italian</cell><cell cols="6">Prec. 10 docs 0.400 0.288 0.444 0.384</cell></row><row><cell></cell><cell cols="6">Av Precision 0.366 0.288 0.438 0.351</cell></row><row><cell></cell><cell cols="2">Rel. Ret.</cell><cell>633</cell><cell>591</cell><cell>602</cell><cell>639</cell></row><row><cell>Spanish</cell><cell cols="6">Prec. 10 docs 0.484 0.316 0.460 0.448</cell></row><row><cell></cell><cell cols="6">Av Precision 0.444 0.318 0.406 0.398</cell></row><row><cell></cell><cell cols="2">Rel. Ret.</cell><cell>767</cell><cell>666</cell><cell>649</cell><cell>755</cell></row><row><cell cols="7">Spanish (rev.) Prec. 10 docs 0.532 0.320 0.492 0.488</cell></row><row><cell></cell><cell cols="6">Av Precision 0.472 0.312 0.410 0.446</cell></row><row><cell></cell><cell cols="2">Rel. Ret.</cell><cell>775</cell><cell>657</cell><cell>647</cell><cell>774</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2 :</head><label>2</label><figDesc>Retrieval runs with PRF.</figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 3 :</head><label>3</label><figDesc>Retrieval runs fusing PRF runs with standard VIPER image matching results.</figDesc><table><row><cell></cell><cell></cell><cell>SDL</cell><cell>INT</cell><cell>ST</cell><cell>MG</cell></row><row><cell>Dutch</cell><cell cols="5">Prec. 10 docs 0.480 0.276 0.508 0.464</cell></row><row><cell></cell><cell cols="5">Av Precision 0.394 0.273 0.433 0.419</cell></row><row><cell></cell><cell>Rel. Ret.</cell><cell>638</cell><cell>637</cell><cell>709</cell><cell>791</cell></row><row><cell>French</cell><cell cols="5">Prec. 10 docs 0.472 0.520 0.496 0.428</cell></row><row><cell></cell><cell cols="5">Av Precision 0.407 0.466 0.428 0.399</cell></row><row><cell></cell><cell>Rel. Ret.</cell><cell>666</cell><cell>707</cell><cell>658</cell><cell>695</cell></row><row><cell cols="6">German Prec. 10 docs 0.604 0.524 0.548 0.636</cell></row><row><cell></cell><cell cols="5">Av Precision 0.501 0.467 0.474 0.532</cell></row><row><cell></cell><cell>Rel. Ret.</cell><cell>763</cell><cell>804</cell><cell>691</cell><cell>804</cell></row><row><cell>Italian</cell><cell cols="5">Prec. 10 docs 0.400 0.288 0.440 0.392</cell></row><row><cell></cell><cell cols="5">Av Precision 0.369 0.289 0.437 0.351</cell></row><row><cell></cell><cell>Rel. Ret.</cell><cell>633</cell><cell>591</cell><cell>602</cell><cell>639</cell></row><row><cell cols="6">Spanish Prec. 10 docs 0.472 0.324 0.452 0.444</cell></row><row><cell></cell><cell cols="5">Av Precision 0.441 0.316 0.405 0.397</cell></row><row><cell></cell><cell>Rel. Ret.</cell><cell>767</cell><cell>666</cell><cell>649</cell><cell>755</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 4 :</head><label>4</label><figDesc>Retrieval runs with pseudo relevance feedback.</figDesc><table><row><cell></cell><cell></cell><cell>SDL</cell><cell>INT</cell><cell>ST</cell><cell>MG</cell></row><row><cell>Dutch</cell><cell cols="5">Prec. 10 docs 0.116 0.172 0.124 0.140</cell></row><row><cell></cell><cell cols="5">Av Precision 0.105 0.127 0.141 0.121</cell></row><row><cell></cell><cell>Rel. Ret.</cell><cell>638</cell><cell>637</cell><cell>709</cell><cell>791</cell></row><row><cell>French</cell><cell cols="5">Prec. 10 docs 0.128 0.120 0.128 0.112</cell></row><row><cell></cell><cell cols="5">Av Precision 0.107 0.110 0.117 0.100</cell></row><row><cell></cell><cell>Rel. Ret.</cell><cell>666</cell><cell>707</cell><cell>658</cell><cell>695</cell></row><row><cell cols="6">German Prec. 10 docs 0.164 0.172 0.124 0.148</cell></row><row><cell></cell><cell cols="5">Av Precision 0.146 0.169 0.132 0.148</cell></row><row><cell></cell><cell>Rel. Ret.</cell><cell>763</cell><cell>804</cell><cell>691</cell><cell>804</cell></row><row><cell>Italian</cell><cell cols="5">Prec. 10 docs 0.132 0.132 0.140 0.112</cell></row><row><cell></cell><cell cols="5">Av Precision 0.132 0.119 0.118 0.108</cell></row><row><cell></cell><cell>Rel. Ret.</cell><cell>633</cell><cell>591</cell><cell>602</cell><cell>639</cell></row><row><cell cols="6">Spanish Prec. 10 docs 0.140 0.140 0.140 0.132</cell></row><row><cell></cell><cell cols="5">Av Precision 0.145 0.111 0.128 0.131</cell></row><row><cell></cell><cell>Rel. Ret.</cell><cell>767</cell><cell>666</cell><cell>649</cell><cell>755</cell></row></table></figure>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Exeter at CLEF 2003: Experiments with Machine Translation for Monolingual, Bilingual and Multilingual Retrieval</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">M</forename><surname>Lam-Adesina</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">J F</forename><surname>Jones</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of Workshop of the Cross-Language Evaluation Forum (CLEF 2003</title>
				<editor>
			<persName><forename type="first">C</forename><surname>Peters</surname></persName>
		</editor>
		<meeting>Workshop of the Cross-Language Evaluation Forum (CLEF 2003<address><addrLine>Trondheim, Norway</addrLine></address></meeting>
		<imprint>
			<publisher>Springer-Verlag</publisher>
			<date type="published" when="2004">2004</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Applying Summarization Techniques for Term Selection in Relevance Feedback</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">M</forename><surname>Lam-Adesina</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">J F</forename><surname>Jones</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 24th Annual International ACM SIGIR</title>
				<meeting>the 24th Annual International ACM SIGIR<address><addrLine>New Orleans</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2001">2001</date>
			<biblScope unit="page" from="1" to="9" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Bleu: a method for automatic evaluation of machine translation</title>
		<author>
			<persName><forename type="first">K</forename><surname>Papineni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Roukos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Ward</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W.-J</forename><surname>Zhu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 40th Annual Meeting of the Association for Computational Linguistics</title>
				<meeting>the 40th Annual Meeting of the Association for Computational Linguistics<address><addrLine>Philadelphia, USA</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2002">2002. 2002</date>
			<biblScope unit="page" from="311" to="318" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Automatic Evaluation of Machine Translation Quality Using N-gram Co-Occurrence Statistics</title>
		<author>
			<persName><forename type="first">G</forename><surname>Doddington</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Human Language Technology: Notebook Proceedings</title>
				<meeting><address><addrLine>San Diego</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2002">2002</date>
			<biblScope unit="page" from="128" to="132" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<monogr>
		<ptr target="http://nlp.cs.nyu.edu/GTM/" />
		<title level="m">General Text Matcher</title>
				<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<monogr>
		<ptr target="http://www.nist.gov/speech/tests/mt/resources/scoring.htm" />
		<title level="m">NIST&apos;s MT Evaluation Toolkit</title>
				<imprint/>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
