=Paper= {{Paper |id=Vol-1172/CLEF2006wn-adhoc-Monz2006 |storemode=property |title=Statistical Machine Translation and Cross-Language IR: QMUL at CLEF 2006 |pdfUrl=https://ceur-ws.org/Vol-1172/CLEF2006wn-adhoc-Monz2006.pdf |volume=Vol-1172 |dblpUrl=https://dblp.org/rec/conf/clef/Monz06 }} ==Statistical Machine Translation and Cross-Language IR: QMUL at CLEF 2006== https://ceur-ws.org/Vol-1172/CLEF2006wn-adhoc-Monz2006.pdf
Statistical Machine Translation and Cross-Language IR:
                 QMUL at CLEF 2006
                                             Christof Monz
                                    Department of Computer Science
                                    Queen Mary, University of London
                                          London E1 4NS, UK
                                   Email: christof@dcs.qmul.ac.uk
                             Web: http://www.dcs.qmul.ac.uk/∼christof


                                                   Abstract
       In this year’s CLEF submissions we focus on using a state-of-the-art statistical machine trans-
       lation approach for ad-hoc cross-language retrieval. Our machine translation approach is
       phrase-based as opposed to statistical word-based approaches that have been previously used
       for query translation in cross-language IR. The phrase translation probabilities were estimated
       by using the Europarl corpus. For query formulation, we also use the n-best lists of translation
       candidates to assign weights to query terms. Our results show that a statistical phrase-based ap-
       proach is a competitive alternative to commercial, rule-based machine translation approaches
       in the context of cross-language IR.

Categories and Subject Descriptors
H.3 [Information Storage and Retrieval]: H.3.1 Content Analysis and Indexing; H.3.3 Information
Search and Retrieval; H.3.4 Systems and Software; H.3.7 Digital Libraries; H.2.3 [Database Manage-
ment]: Languages—Query Languages

General Terms
Information Retrieval, Performance, Experimentation

Keywords
Cross-Language Information Retrieval, Statistical Machine Translations


1     Introduction
The two basic approaches to cross-language retrieval are: (1) Automatically translate all documents in
the collection into the source language and then apply monolingual retrieval on the translated document
collection; and (2) Automatically translate the user-posed query into the target language and then apply
monolingual retrieval with the translated query on the original document collection in the target language.
    Query translation requires access to some form of translation dictionary. Three approaches may be
used to produce the translations:
    1. Application of a machine translation system to translate the entire query into the target language.
    2. Use of a dictionary to produce a number of target-language translations for words or phrases in the
       source language.
    3. Use of a parallel corpus to estimate the probabilities that word w in the source language translates
       into word w0 in the target language.
The main shortcoming of most commercial machine translation systems is that they only return the most
likely translation, where ‘most likely’ is defined in terms of the internal algorithm of the translation system.
But this does not mean that there are no other equally good—or, by some other objective standard, maybe
even better—alternative translations.
    On the other hand, there are several drawbacks to the parallel-corpus approach as well. First, although
parallel corpora are available for many of the European languages, there are many languages for which there
are still no parallel corpora large enough to estimate translation probabilities. Second, most of the parallel
corpora belong to a rather specific domain, such as the Europarl corpus,1 which contains the proceedings
of the European parliament in 11 languages for the years 1996–2003. This introduces a bias toward the
domain of the parallel corpus and makes the learned translation probabilities less reliable for other domains.
A third disadvantage is that the translation probabilities induced from parallel corpora are typically based
on single-word mappings, although recent template-based statistical methods facilitate the acquisition of
phrase translations [12].
    Here, we propose an approach that uses a state-of-the-art phrase based statistical machine translation
system. Since this system is trained on the aforementioned Europarl corpus, we wanted to investigate
whether such a translation approach results in a well-performing cross-language IR system despite the
difference in genre between the corpus that is used to train the translation model and the retrieval corpus.
    This paper is organized as follows: The next section provides some background on statistical machine
translation and its application to cross-language IR. Section 3 describes the experimental set-up and results
of our runs submitted to CLEF 2006. Finally we draw some conclusions in Section 4.


2      Statistical Machine Translation
This section provides some background on statistical machine translation and how it is applied to cross-
language retrieval.

2.1     Phrase-Based Translation
Word-based statistical machine translation approaches have been used in several cross-language IR sys-
tems, see, e.g. [3, 7, 8]. These approaches learn word translations probabilities from a parallel corpus, or
bi-text, which is a set of sentence pairs, where both sentences in a pair are translations of each other. There
are a number of approaches computing these translation probabilities, such as the IBM models [1], Hidden
Markov Models [18] and other co-occurrence measures. In practice most cross-language IR systems use
IBM model 1, the most the simple of the IBM models.
    In the area of machine translation word-based translation approaches more recently have been replaced
by phrase-bases approaches, see, e.g. [12], which generate translations of substantially better quality.
Phrase-based models use translation-probabilities for sequences of consecutive words—although not nec-
essarily linguistic phrases—instead of individual words. The advantage is that more contextual information
is captured this way, and at the same time this approach models some of the local word re-orderings that
can take place when translating from one language into another. Word re-ordering does not appear to be a
prime concern in cross-language retrieval, where most systems use a bag-of-words approach, and fluency
of the translation can be disregarded. On the other hand, aiming to generate fluent output can also have a
positive impact on choosing a correct phrase translation. In general, the probability of the English sentence
e being a translation of the foreign sentence f is computed by applying Bayes’ rule:

                                                             p( f |e) · p(e)
                                              p(e| f ) =                                                    (1)
                                                                  p( f )
    1 The Europarl corpus is freely available from http://www.isi.edu/∼koehn/europarl/.
Since we are only interested in finding the most likely English translation for a given foreign sentence,
p( f ) can be disregarded:

                                 argmaxe p(e| f ) = argmaxe p( f |e) · p(e)                                     (2)

Here, the likelihood, p(e| f ), models the faithfulness of translation, i.e. to what extent it covers the semantic
content of the foreign sentence, and the prior, p(e), models the fluency of the generated translation.
    During decoding, which is the process of finding the English sentence maximizing p(e| f ) according
to equation (2), English translation candidates are generated incrementally by applying matching phrase
translations and considering the probability of the fluency of each translation candidate. A number of
decoding approaches are used in statistical machine translation, and here we follow the multi-stack beam
search design used in [5].
    In multi-stack beam search decoding partially generated translation candidates are grouped in stacks,
where candidates are put into the same stack if they have translated the same number of words from the
foreign sentence. Since comprehensive search is computationally prohibitive, and in fact NP-complete, see
[4], heuristics have to be used to restrict the search space. See [5] for more details on multi-stack decoding.

2.2    Phrase Extraction
Although most alignment approaches, such as the IBM models 1–5 [1], are restricted to word-level align-
ments, it is possible to extract phrase translations by post-processing the word-level alignments, and there
are a number of approaches to phrase extraction, see, e.g., [11, 17]. For our experiments we use the ap-
proach described in [6], which is also part of the distribution of the Pharaoh Statistical Machine Translation
system.2
    In this approach, first the intersection of the alignment links of both alignment directions, i.e. e-to- f
and f -to-e, is taken. This initial alignment is then expanded by adding links that are adjacent to existing
links, which is done iteratively until no further alignment links can be added.
    During phrase extraction, groups of adjacent words are extracted on the foreign and the English side.
Two phrases are considered translations of each other if the words in the phrase pair are only aligned to
each other and not to words outside the phrase pair. The conditional phrase translation probabilities are
simply computed by marginalizing the joint distribution of the phrase pair:

                                                               freq( f¯, ē)
                                                p( f¯|ē) =
                                                              ∑ f¯ freq( f¯, ē)

where f¯ is a phrase in the foreign language, and ē is an English phrase. For more details see [5].

2.3    Language Modeling
When estimating the probability of a translation candidate, it is impossible to use maximum likelihood
estimates on the entire string, by using the chain rule

                       p(e) =       p(e1 ) · p(e2 |e1 ) · p(e3 |e1 , e2 ) · . . . · p(en |e1 , . . . , en−1 )   (3)

where e is a string in target language of length n, as one would run into serious data sparseness issues.
Instead most language modeling approaches use a limited horizon of two words, called tri-gram models:

                        p(e) ≈       p(e1 ) · p(e2 |e1 ) · p(e3 |e1 , e2 ) · . . . · p(en |en−2 , en−1 )        (4)

Despite the limited context that is taken into account they still form a good compromise between prediction
quality, robustness and computational tractability.
  2 Pharaoh is freely available from http://www.isi.edu/licensed-sw/pharaoh/.
2.4     Parameter Estimation
During decoding the different probabilities are combined by using a log-linear model. In addition to the
translation and language model probabilities our model contains parameters that regulate, for example,
word re-ordering, length of the translation, and number of phrase matches. Manually estimating the appro-
priate weights for all parameters is nearly impossible as the number of combinations is simply too large.
Hence some form of optimization is required. Here we use the minimum error rate training (MERT) ap-
proach introduced by Franz Och [10].3 Although MERT is not guaranteed to find a global optimum, its
parameter estimations generally lead to substantial improvements in translation quality.

2.5     Translating Retrieval Queries
As mentioned above, full-fledged machine translation systems expect their input to be well-formed lin-
guistic units, normally sentences. Statistical machine translation systems are no exception, but it should
be pointed out that this restriction does not hold in principle, but is mainly due to the way statistical MT
systems are trained.
     When translating the queries, no stop word removal is applied. The only pre-processing that is applied
is tokenization and case folding, as the entries in the phrase table are case-folded as well. The main reason
for keeping stop words is that we want to keep the input to the MT system fluent.
     One of the major disadvantages of using commercial machine translation systems is that they tend to
return only the most likely translation, despite the fact that there are many more translations that might
be almost as good and formulate the translation slightly different. In addition, if the machine translation
system is mainly rule-based, scoring different translation candidates can be rather difficult as it is hard to
consistently assign weights to human-generated translation rules.
     On the other hand it is very easy to generate n-best list for statistical machine translation systems, as
all translation and language model probabilities are estimated by using training data.


3      CLEF 2006 Experiments
This year, we participated in the English-to-French and English-to-Portuguese cross-language retrieval
tasks. The following three subsections describe the configurations and data used for our machine translation
component and the information retrieval system, and report the official CLEF 2006 results for those runs.

3.1     Experimental Set-Up: Machine Translation
For both translation directions we used the Europarl parallel corpus to estimate the phrase translation
probabilities. Table 1 summarizes some collection statistics of the data used to build the machine translation
systems for our cross-language retrieval experiments.

                                                                   English-French         English-Portuguese
              no. of sentence pairs                                       645,518                    568,446
              no. of English words                                          14.4M                      13.2M
              no. of foreign words                                          16.3M                      13.5M
              no. of distinct phrase pairs                                  28.0M                      29.3M
              no. of distinct phrase pairs (CLEF 2006)                       0.6M                       0.6M
              no. of words in language model corpus                         87.5M                    126.4M
              no. of distinct 1-grams                                      512.9K                     509.7K
              no. of distinct 2-grams                                        5.8M                      10.3M
              no. of distinct 3-grams                                        5.8M                      10.3M

                          Table 1: Overview of the data used in the statistical MT system.

    3 We are grateful to Philipp Koehn who provided us with a re-implementation of Och’s optimization procedure.
    The first five rows in Table 1 refer to parallel corpora from Europarl that were used to estimate phrase
translation probabilities. The resulting phrase translation table is rather big and hard to keep in memory,
and it was filtered to retain only phrases that match word sequences occurring in the CLEF 2006 topic set
(see row 5).
    In order to build the French and the Portuguese language models, we used the French and Portuguese
side of the parallel corpus in combination with the documents in the CLEF retrieval corpus: ATS 1994–
1995 and Le Monde 1994–1995 for French, and Fohla 1994–1995 and Publico 1994–1995 for Portuguese,
see Table 1 for the respective sizes. We used the SRI Language Modeling toolkit [16] to train the tri-gram
model, using modified Kneser-Ney smoothing [2]. The distribution of n-grams is shown in the last three
rows of Table 1.
    The parallel corpora were tokenized and case-folded. We used the simple rule-based tokenization
scripts for all three languages involved. Diacritics were not removed.
    For query translation we used the title and description fields only, and each field was translated sep-
arately. Note that in Section 2 we assumed that we translate from a foreign language into English, as
most MT approaches focus on translation into English. For our experiments we translate from English into
French and Portuguese, respectively. Hence English is referred to as the foreign language in the discussion
below.
    Parameter estimation for the translation models used CLEF test sets from previous years. Most of
the CLEF topics are available in several languages where topics have been created by a human translator.
For English-to-French translation we used the topics 210 from CLEF 2001–2004, and for English-to-
Portuguese translation we used the 100 topics from 2002 and 2004.

3.2   Experimental Set-Up: Information Retrieval
The submitted runs used F LEX IR, a home-grown information retrieval system [9]. The main goal under-
lying F LEX IR’s design is to facilitate flexible experimentation with a wide variety of retrieval components
and techniques. F LEX IR is implemented in Perl; as it is built around the standard UNIX pipeline architec-
ture, and supports many types of preprocessing, scoring, indexing, and retrieval tools. The retrieval model
underlying F LEX IR is the standard vector space model. The Okapi BM25 weighting scheme [13] is used
to compute the similarity between a query and a document. The BM25 parameters were set to k1 = 1.5,
k3 = 7, and b = 0.7 as suggested in [15].
    The text collection files were pre-processed using the same French and Portuguese tokenizers that were
used to pre-process the parallel corpus. Stop words were removed from the translated queries, using the
stop word list provided by the University of Neuchatel.4
    Stemming and n-gram splitting was used for the French run, where we used the S NOW BALL stemmer,
which was provided by the University of Neuchatel as well. After stemming, n-gram splitting was applied
with n = 4. Note that n-gram splitting, was only applied to individual tokens and did not cross word
boundaries.
    The Portuguese run used only ngram splitting and no stemming. Here n = 5 and n-grams did cross word
boundaries. For both French and Portuguese, these settings were determined by using the CLEF 2005 test
collection as a development set.
    Blind relevance feedback was applied to expand the original query with related terms. Term weights
were computed by using the standard Rocchio method [14], where we considered the top 10 documents to
be relevant and the bottom 500 documents to be non-relevant. We allowed at most 20 terms to be added to
the original query.
    As we mentioned above, both topic fields, i.e., the title and the description, were translated separately.
For retrieval both translations were merged into a single query. For our official submissions we also used
the n-best lists generated by our statistical machine translation system. The n-best lists are used to assign
weights to terms in the query:

                                                     ∑ni=1 occurs in(t, si ) · p(si | f )
                                 weight f (t) =                                                           (5)
                                                          ∑ni=1 ∑ t∈si p(si | f )
  4 The stop word lists are available at http://www.unine.ch/info/clef/.
                                           English-to-French
                                         QMUL06e2f10b 0.3396
                                          English-to-Portuguese
                                         QMUL06e2p10b 0.3526

                                     Table 2: Official CLEF 2006 results



where si is the i-th translation candidate for f and f is the original foreign language topic (note: here,
English is the foreign language as we translate from English into French and Portuguese). We experimented
with n-best lists of different length, and setting n = 10 yielded the best results for the CLEF 2005 test set.
The term weight in (5) is computed for each occurrence of a term in a query.

3.3    CLEF 2006 Experimental Results
We submitted two official runs to CLEF 2006, one for the English-to-French cross-lingual task and one
for the English-to-Portuguese cross-lingual task. The MAP scores for both runs are shown in Table 2.
Although using n-best machine translation output for term re-weighting did lead to slight improvements
for the CLEF 2005 development set, the improvements were still rather small. In part this could be due to
the small variation in the n-best lists. The differences between the translation candidates tend to be minute
and are often causes by different derivations of almost identical translations. For example, Table 3 shows
the top ten translations for topic 301.

             Rank                                  Translation
                       English: what brands are marketed by nestle around the world ?
                  1    les marques sont commercialisees par nestle dans le monde ?
                  2    trans=ce que sont marques commercialisees par nestle dans le monde ?
                  3    les marques sont commercialisees par nestle dans le monde - ?
                  4    ce que marques sont commercialisees par nestle dans le monde ?
                  5    ce que sont marques commercialisees par nestle dans le monde - ?
                  6    ce que marques sont commercialisees par nestle dans le monde - ?
                  7    les marques sont commercialisees par nestle dans le monde ?
                  8    ce qu ’ marques sont commercialisees par nestle dans le monde ?
                  9    ce qui marques sont commercialisees par nestle dans le monde ?
                 10    les marques sont commercialisees par nestle dans le monde entier ?

                                 Table 3: Top ten translations for topic 301


    As one can see, the translation candidates are almost identical to each other, and there is little variation
in terms of lexical translation choices. Using longer n-best lists did not change this fact. It seems that
more elaborate selection strategies are needed to filter out translation candidates that are merely due to
word-order or phrase-matching differences.


4     Conclusions
With our submissions we tried investigate to what extent statistical machine translation is effective in
the context of cross-language information retrieval. In particular with respect to machine translation ap-
proaches that translate the entire topic as opposed to word-translation models like the simpler IBM models.
One of the main advantages of statistical machine translation is the ease with which n-best translation out-
put can be generated, but unfortunately we were only able to get minor improvements from this. More
elaborate ways of exploiting n-best lists remains a topic that we want to pursue further in future research.
Acknowledgments
This work has been supported, in part, by the Nuffield Foundation, Grant No. NAL/32720.


References
 [1] Peter F. Brown, Stephen A. Della Pietra, Vincent J. Della Pietra, and Robert L. Mercer. The mathemat-
     ics of statistical machine translation: parameter estimation. Computational Linguistics, 19(2):263–
     311, 1993.

 [2] Stanley Chen and Joshua Goodman. An empirical study of smoothing techniques for language mod-
     eling. Technical Report TR-10-98, Harvard University, 1998.
 [3] Martin Franz, J. Scott McCarley, Todd Ward, and Wei-Jing Zhu. Quantifying the utility of parallel
     corpora. In Proceedings of the 24th Annual International ACM SIGIR Conference on Research and
     Development in Information Retrieval, pages 398–399, 2001.

 [4] Kevin Knight. Decoding complexity in word-replacement translation models. Computational Lin-
     guistics, 25(4):607–615, 1999.
 [5] Philipp Koehn. Pharaoh: A beam search decoder for phrase-based statistical machine translation
     models. In Robert E. Frederking and Kathryn B. Taylor, editors, Proceedings of the 6th Conference
     of the Association for Machine Translations in the Americas (AMTA 2004), pages 115–124, 2004.
 [6] Philipp Koehn. Pharaoh: A beam search decoder for phrase-based statistical machine translation
     models, user manual. Technical report, USC Information Science Institute, 2004.
 [7] Wessel Kraaij, Jian-Yun Nie, and Michel Simard. Embedding web-based statistical translation models
     in cross-language information retrieval. Computational Linguistics, 29(3):381–419, 2003.
 [8] Paul McNamee. Exploring new languages with HAIRCUT at CLEF 2005. In Proceedings of the 6th
     Workshop of the Cross-Language Evaluation Forum (CLEF 2005), 2005.
 [9] Christof Monz, Jaap Kamps, and Maarten de Rijke. The University of Amsterdam at CLEF 2002. In
     Working Notes for the CLEF 2002 Workshop, pages 73–84, 2002.
[10] Franz-Josef Och. Minimum error rate training in statistical machine translation. In Proceedings of the
     41st Annual Meeting of the Association for Computational Linguistics (ACL), pages 160–167, 2003.
[11] Franz Josef Och and Hermann Ney. A systematic comparison of various statistical alignment models.
     Computational Linguistics, 29(1):19–52, 2003.
[12] Franz-Josef Och and Hermann Ney. The alignment template approach to statistical machine transla-
     tion. Computational Linguistics, 30(4):417–449, 2004.
[13] Stephen E. Robertson, Steve Walker, Micheline Hancock-Beaulieu, Aarron Gull, and Marianna Lau.
     Okapi at TREC-3. In Donna K. Harman, editor, Proceedings of the Third Text Retrieval Conference
     (TREC-3), pages 109–126. NIST Special Publication 500-225, 1994.
[14] Joseph J. Rocchio. Relevance feedback in information retrieval. In G. Salton, editor, The SMART
     Retrieval System — Experiments in Automatic Document Processing, pages 313–323. Prentice Hall,
     1971.
[15] Jacques Savoy and Pierre-Yves Berger. Report on CLEF-2005 evaluation campaign: Monolingual,
     bilingual, and GIRT information retrieval. In Carol Peters, Paul Clough, Julio Gonzalo, Gareth J.F.
     Jones, Michael Kluck, and Bernardo Magnini, editors, Proceedings of the 6th Workshop of the Cross-
     Language Evaluation Forum (CLEF 2005), 2005.
[16] Andreas Stolcke. SRILM—an extensible language modeling toolkit. In Proceedings of the Interna-
     tional Conference on Spoken Language Processing, pages 901–904, 2002.
[17] Ashish Venugopal, Stephan Vogel, and Alex Waibel. Effective phrase translation extraction from
     alignment models. In Proceedings of the 41st Annual Meeting of the Association for Computational
     Linguistics (ACL-2003), pages 319–326, 2003.

[18] Stephan Vogel, Hermann Ney, and Christoph Tillmann. HMM-based word alignment in statistical
     translation. In Proceedings of the 16th conference on Computational Linguistics (COLING ’96),
     pages 836–841, 1996.