=Paper= {{Paper |id=Vol-1175/CLEF2009wn-adhoc-AgirreEt2009a |storemode=property |title=CLEF 2009 Ad Hoc Track Overview: Robust - WSD Task |pdfUrl=https://ceur-ws.org/Vol-1175/CLEF2009wn-adhoc-AgirreEt2009a.pdf |volume=Vol-1175 |dblpUrl=https://dblp.org/rec/conf/clef/AgirreNMO09a }} ==CLEF 2009 Ad Hoc Track Overview: Robust - WSD Task== https://ceur-ws.org/Vol-1175/CLEF2009wn-adhoc-AgirreEt2009a.pdf
            CLEF 2009 Ad Hoc Track Overview:
                   Robust-WSD Task

          Eneko Agirre1 , Giorgio Maria Di Nunzio2 , Thomas Mandl3 , and
                                  Arantxa Otegi1
     1
         Computer Science Department, University of the Basque Country, Spain
                        {e.agirre,arantza.otegi}@ehu.es
         2
           Department of Information Engineering, University of Padua, Italy
                             {dinunzio}@dei.unipd.it
             3
               Information Science, University of Hildesheim, Germany
                             mandl@uni-hildesheim.de


         Abstract. The Robust-WSD at CLEF 2009 aims at exploring the con-
         tribution of Word Sense Disambiguation to monolingual and multilingual
         Information Retrieval. The organizers of the task provide documents
         and topics which have been automatically tagged with Word Senses
         from WordNet using several state-of-the-art Word Sense Disambigua-
         tion systems. The Robust-WSD exercise follows the same design as in
         2008. It uses two languages often used in previous CLEF campaigns (En-
         glish, Spanish). Documents were in English, and topics in both English
         and Spanish. The document collections are based on the widely used
         LA94 and GH95 news collections. All instructions and datasets required
         to replicate the experiment are available from the organizers website
         (http://ixa2.si.ehu.es/clirwsd/). The results show that some top-scoring
         systems improve their IR and CLIR results with the use of WSD tags,
         but the best scoring runs do not use WSD.


1    Introduction
The Robust-WSD task at CLEF 2009 aims at exploring the contribution of
Word Sense Disambiguation to monolingual and multilingual Information Re-
trieval. The organizers of the task provide documents and topics which have
been automatically tagged with Word Senses from WordNet using several state-
of-the-art Word Sense Disambiguation systems. The task follows the same design
as in 2008.
    The robust task ran for the fourth time at CLEF 2009. It is an Ad-Hoc
retrieval task based on data of previous CLEF campaigns. The robust task em-
phasizes the difficult topics by a non-linear integration of the results of individual
topics into one result for a system, using the geometric mean of the average pre-
cision for all topics (GMAP) as an additional evaluation measure [13,14]. Given
the difficulty of the task, training data including topics and relevance assessments
was provided for the participants to tune their systems to the collection.
    For the second year, the robust task also incorporated word sense disambigua-
tion information provided by the organizers to the participants. The task follows
the 2007 joint SemEval-CLEF task [2] and the 2008 Robust-WSD exercise [3],
and has the aim of exploring the contribution of word sense disambiguation to
monolingual and cross-language information retrieval. The goal of the task is
to test whether WSD can be used beneficially for retrieval systems, and thus
participants were required to submit at least one baseline run without WSD and
one run using the WSD annotations. Participants could also submit four further
baseline runs without WSD and four runs using WSD.
    The experiment involved both monolingual (topics and documents in En-
glish) and bilingual experiments (topics in Spanish and documents in English).
In addition to the original documents and topics, the organizers of the task pro-
vided both documents and topics which had been automatically tagged with
word senses from WordNet version 1.6 using two state-of-the-art word sense dis-
ambiguation systems, UBC [1] and NUS [7]. These systems provided weighted
word sense tags for each of the nouns, verbs, adjectives and adverbs that they
could disambiguate.
    In addition, the participants could use publicly available data from the En-
glish and Spanish wordnets in order to test different expansion strategies. Note
that given the tight alignment of the Spanish and English wordnets, the wordnets
could also be used to translate directly from one sense to another, and perform
expansion to terms in another language.
    The datasets used in this task can be used in the future to run further
experiments. Check http://ixa2.si.ehu.es/clirwsd for information of how
to access the datasets. Topics and relevance judgements are freely available.
The document collection can be obtained from ELDA purchasing the CLEF
Test Suite for the CLEF 2000-2003 Campaigns – Evaluation Package. As an
alternative, the website offers the unordered set of words in each document,
that is, the full set of documents where the positional information has been
eliminated to avoid replications of the originals. Lucene indexes for the later are
also available from the website.
    In this paper, we first present the task setup, the evaluation methodology and
the participation in the different tasks (Section 2). We then describe the main
features of each task and show the results (Sections 3 - 5). The final section
provides a brief summing up. For information on the various approaches and
resources used by the groups participating in this task and the issues they focused
on, we refer the reader to the rest of the papers in the Robust-WSD part of the
Ad Hoc section of these Proceedings.


2   Task Setup

The Ad Hoc task in CLEF adopts a corpus-based, automatic scoring method
for the assessment of system performance, based on ideas first introduced in
the Cranfield experiments in the late 1960s [8]. The tasks offered are studied
in order to effectively measure textual document retrieval under specific condi-
tions. The test collections are made up of documents, topics and relevance
assessments. The topics consist of a set of statements simulating information
needs from which the systems derive the queries to search the document collec-
tions. Evaluation of system performance is then done by judging the documents
retrieved in response to a topic with respect to their relevance, and computing
the recall and precision measures.

2.1    Test Collections
The Documents. The robust task used existing CLEF news collections but
with word sense disambiguation (WSD) information added. The word sense dis-
ambiguation data was automatically added by systems from two leading research
laboratories, UBC [1] and NUS [7]. Both systems returned word senses from the
English WordNet, version 1.6.
    The document collections were offered both with and without WSD, and
included the following1 :
 – LA Times 94 (with word sense disambiguated data); ca 113,000 documents,
   425 MB without WSD, 1,448 MB (UBC) or 2,151 MB (NUS) with WSD;
 – Glasgow Herald 95 (with word sense disambiguated data); ca 56,500 doc-
   uments, 154 MB without WSD, 626 MB (UBC) or 904 MB (NUS) with
   WSD.

The Topics. Topics are structured statements representing information needs.
Each topic typically consists of three parts: a brief title statement; a one-sentence
description; a more complex narrative the relevance assessment criteria. Topics
are prepared in xml format and identified by means of a Digital Object Identifier
(DOI)2 of the experiment [12] which allows us to reference and cite them.
    The WSD robust task used existing CLEF topics in English and Spanish as
follows:

 – CLEF 2001; Topics 10.2452/41-AH – 10.2452/90-AH; LA Times 94
 – CLEF 2002; Topics 10.2452/91-AH – 10.2452/140-AH; LA Times 94
 – CLEF 2003; Topics 10.2452/141-AH – 10.2452/200-AH; LA Times 94, Glas-
   gow Herald 95
 – CLEF 2004; Topics 10.2452/201-AH – 10.2452/250-AH; Glasgow Herald 95
 – CLEF 2005; Topics 10.2452/251-AH – 10.2452/300-AH; LA Times 94, Glas-
   gow Herald 95
 – CLEF 2006; Topics 10.2452/301-AH – 10.2452/350-AH; LA Times 94, Glas-
   gow Herald 95

    Topics from years 2001, 2002 and 2004 were used as training topics (relevance
assessments were offered to participants), and topics from years 2003, 2005 and
2006 were used for the test.
    All topics were offered both with and without WSD. Topics in English were
disambiguated by both UBC [1] and NUS [7] systems, yielding word senses from
1
    A sample document and dtd are available at http://ixa2.si.ehu.es/clirwsd/
2
    http://www.doi.org/
                1/1
                file:/Users/ferro/Documents/Pubblicazioni/2008/CLEF/WN/ad-hoc/figures/topic_141-WSD-AH.xml

               
                   10.2452/141-WSD-AH

                       
                           
                               Letter
                               
                               
                               
                               
                           

                               
                                   Bomb
                                   
                                   
                                   
                               

                               
                                   for
                               

                               ...

                       

                       
                           
                               Find
                               

                                        ...

                               

                               ...

                       

                   
                       ...
                   
               



          Fig. 1. Example of Robust WSD topic: topic 10.2452/141-WSD-AH.



WordNet version 1.6. A large-scale disambiguation system for Spanish was not
available, so we used the first-sense heuristic, yielding senses from the Spanish
wordnet, which is tightly aligned to the English WordNet version 1.6 (i.e., they
share synset numbers or sense codes). An excerpt from a topic is shown in Figure
1, where each term in the topic is followed by its senses with their respective
scores as assigned buy the automatic WSD system3 .


Relevance Assessment. The number of documents in large test collections
such as CLEF makes it impractical to judge every document for relevance. In-
stead approximate recall values are calculated using pooling techniques. The
robust WSD task used existing relevance assessments from previous years. The
3
    Full sample and dtd are available at http://ixa2.si.ehu.es/clirwsd/
relevance assessments regarding the training topics were provided to participants
before competition time.
    The total number of assessments was 66,441 documents of which 4,327 were
relevant. The distribution of the pool according to each year was the following:

 – CLEF 2003: 23,674 documents, 1,006 relevant;
 – CLEF 2005: 19,790 document, 2,063 relevant;
 – CLEF 2006: 21,247 document, 1,258 relevant;

    Seven topics had no relevant documents at all: 10.2452/149-AH, 10.2452/161-
AH, 10.2452/166-AH, 10.2452/186-AH, 10.2452/191-AH, 10.2452/195-AH, 10.2-
452/321-AH. Each topic had an average of about 28 relevant documents and a
standard deviation of 34, a minimum of 1 relevant document and a maximum
of 229 relevant documents per topic.


2.2   Result Calculation

Evaluation campaigns such as TREC and CLEF are based on the belief that
the effectiveness of Information Retrieval Systems (IRSs) can be objectively
evaluated by an analysis of a representative set of sample search results. For
this, effectiveness measures are calculated based on the results submitted by the
participants and the relevance assessments. Popular measures usually adopted
for exercises of this type are Recall and Precision. Details on how they are
calculated for CLEF are given in [6].
    The robust task emphasizes the difficult topics by a non-linear integration of
the results of individual topics into one result for a system, using the geometric
mean of the average precision for all topics (GMAP) as an additional evaluation
measure [13,14].
    The individual results for all official Ad Hoc experiments in CLEF 2009 are
given in the one of the Appendices of the CLEF 2009 Working Notes [9].


2.3   Participants and Experiments

As shown in Table 1, 10 groups submitted 89 runs for the Robust tasks:

 – 8 groups submitted monolingual non-WSD runs (25 runs out of 89);
 – 5 groups also submitted bilingual non-WSD runs (13 runs out of 89).

All groups submitted WSD runs (51 out of 89 runs):

 – 10 groups submitted monolingual WSD runs (33 out of 89 runs)
 – 5 groups submitted bilingual WSD runs (18 out of 89 runs)

    Table 2 provides a breakdown of the number of participants and submitted
runs by task. Note that jaen submitted a monolingual non-WSD run as if it was
a WSD run, and that alicante missed to send their non-WSD run on time. The
figures below are the official figures.
                Table 1. CLEF 2009 Ad Hoc Robust participants

participant task                                  No. experiments
alicante    AH-ROBUST-WSD-MONO-EN-TEST-CLEF2009                 3
darmstadt AH-ROBUST-MONO-EN-TEST-CLEF2009                       5
darmstadt AH-ROBUST-WSD-MONO-EN-TEST-CLEF2009                   5
geneva      AH-ROBUST-MONO-EN-TEST-CLEF2009                     5
geneva      AH-ROBUST-WSD-BILI-X2EN-TEST-CLEF2009               1
geneva      AH-ROBUST-WSD-MONO-EN-TEST-CLEF2009                 2
ixa         AH-ROBUST-BILI-X2EN-TEST-CLEF2009                   1
ixa         AH-ROBUST-MONO-EN-TEST-CLEF2009                     1
ixa         AH-ROBUST-WSD-BILI-X2EN-TEST-CLEF2009               4
ixa         AH-ROBUST-WSD-MONO-EN-TEST-CLEF2009                 3
jaen        AH-ROBUST-WSD-MONO-EN-TEST-CLEF2009                 2
know-center AH-ROBUST-BILI-X2EN-TEST-CLEF2009                   3
know-center AH-ROBUST-MONO-EN-TEST-CLEF2009                     3
know-center AH-ROBUST-WSD-BILI-X2EN-TEST-CLEF2009               3
know-center AH-ROBUST-WSD-MONO-EN-TEST-CLEF2009                 3
reina       AH-ROBUST-BILI-X2EN-TEST-CLEF2009                   5
reina       AH-ROBUST-MONO-EN-TEST-CLEF2009                     5
reina       AH-ROBUST-WSD-BILI-X2EN-TEST-CLEF2009               5
reina       AH-ROBUST-WSD-MONO-EN-TEST-CLEF2009                 5
ufrgs       AH-ROBUST-BILI-X2EN-TEST-CLEF2009                   1
ufrgs       AH-ROBUST-MONO-EN-TEST-CLEF2009                     1
ufrgs       AH-ROBUST-WSD-MONO-EN-TEST-CLEF2009                 1
uniba       AH-ROBUST-BILI-X2EN-TEST-CLEF2009                   3
uniba       AH-ROBUST-MONO-EN-TEST-CLEF2009                     3
uniba       AH-ROBUST-WSD-BILI-X2EN-TEST-CLEF2009               5
uniba       AH-ROBUST-WSD-MONO-EN-TEST-CLEF2009                 5
valencia    AH-ROBUST-MONO-EN-TEST-CLEF2009                     2
valencia    AH-ROBUST-WSD-MONO-EN-TEST-CLEF2009                 4

                       Table 2. Number of runs per track.

                           Track              # Part. # Runs
              Robust Mono English Test              8     25
              Robust Mono English Test WSD         10     33
              Robust Biling. English Test           5     13
              Robust Biling. English Test WSD       5     18



3   Results

Table 3 shows the best results for the monolingual runs, and Table 4 shows
the best results for the bilingual runs. In the following pages, Figures 2 and 3
compare the performances of the best systems in terms of average precision
of the top participants of the Robust Monolingual and Monolingual WSD, and
Figures 4 and 5 compare the performances of the best participants of the Robust
Bilingual and Bilingual WSD.
                              Table 3. Best entries for the robust monolingual task.

       Track Rank Participant                  Experiment DOI                                 MAP GMAP
              1st darmstadt 10.2415/AH-ROBUST-MONO-EN-TEST-CLEF2009.DARMSTADT.DA 4            45.09% 20.42%
              2nd reina       10.2415/AH-ROBUST-MONO-EN-TEST-CLEF2009.REINA.ROB2              44.52% 21.18%
              3rd uniba       10.2415/AH-ROBUST-MONO-EN-TEST-CLEF2009.UNIBA.UNIBAKRF          42.50% 17.93%
      English
              4th geneva      10.2415/AH-ROBUST-MONO-EN-TEST-CLEF2009.GENEVA.ISIENNATTDN      41.71% 17.88%
              5th know-center 10.2415/AH-ROBUST-MONO-EN-TEST-CLEF2009.KNOW-CENTER.ASSO        41.70% 18.64%
              1st darmstadt 10.2415/AH-ROBUST-WSD-MONO-EN-TEST-CLEF2009.DARMSTADT.DA WSD 4 45.00% 20.49%
              2nd uniba       10.2415/AH-ROBUST-WSD-MONO-EN-TEST-CLEF2009.UNIBA.UNIBAKEYSYNRF 43.46% 19.60%
      English 3rd know-center 10.2415/AH-ROBUST-WSD-MONO-EN-TEST-CLEF2009.KNOW-CENTER.ASSOWSD 42.22% 19.47%
       WSD 4th reina          10.2415/AH-ROBUST-WSD-MONO-EN-TEST-CLEF2009.REINA.ROBWSD2       41.23% 18.38%
              5th geneva      10.2415/AH-ROBUST-WSD-MONO-EN-TEST-CLEF2009.GENEVA.ISINUSLWTDN 38.11%  16.26%

                                Table 4. Best entries for the robust bilingual task.

Track Rank Participant                         Experiment DOI                                        MAP GMAP
       1st reina       10.2415/AH-ROBUST-BILI-X2EN-TEST-CLEF2009.REINA.BILI2                         38.42% 15.11%
      2nd uniba        10.2415/AH-ROBUST-BILI-X2EN-TEST-CLEF2009.UNIBA.UNIBACROSSKEYRF               38.09% 13.11%
       3rd know-center 10.2415/AH-ROBUST-BILI-X2EN-TEST-CLEF2009.KNOW-CENTER.BILIASSO                28.98% 06.79%
Es-En
       4th ufrgs       10.2415/AH-ROBUST-BILI-X2EN-TEST-CLEF2009.UFRGS.BILINGUAL                     27.65% 07.37%
       5th ixa         10.2415/AH-ROBUST-BILI-X2EN-TEST-CLEF2009.IXA.ESENNOWSD                       18.05% 01.90%
       1st uniba       10.2415/AH-ROBUST-WSD-BILI-X2EN-TEST-CLEF2009.UNIBA.UNIBACROSSKEYSYNRF        37.53% 13.82%
      2nd geneva       10.2415/AH-ROBUST-WSD-BILI-X2EN-TEST-CLEF2009.GENEVA.ISINUSWSDTD              36.63% 16.02%
Es-En 3rd reina        10.2415/AH-ROBUST-WSD-BILI-X2EN-TEST-CLEF2009.REINA.BILIWSD2                  30.32% 09.38%
       4th know-center 10.2415/AH-ROBUST-WSD-BILI-X2EN-TEST-CLEF2009.KNOW-CENTER.BILIASSOWSD         29.64% 07.05%
WSD 5th ixa            10.2415/AH-ROBUST-WSD-BILI-X2EN-TEST-CLEF2009.IXA.ESEN1STTOPSBESTSENSE500DOCS 18.38% 01.98%



               The comparison of the bilingual runs with respect to the monolingual results
           yield the following:
             – ES → EN: 85.2% of best monolingual English IR system (MAP);
             – ES → EN WSD: 83.3% of best monolingual English IR system (MAP);


           3.1     Statistical Testing
           When the goal is to validate how well results can be expected to hold beyond
           a particular set of queries, statistical testing can help to determine what differ-
           ences between runs appear to be real as opposed to differences that are due to
           sampling issues. We aim to identify whether the results of the runs of a task
           are significantly different from the results of other tasks. In particular, we want
           to test whether there is any difference between applying WSD techniques or
           not. Significantly different in this context means that the difference between the
           performance scores for the runs in question appears greater than what might
           be expected by pure chance. As with all statistical testing, conclusions will be
           qualified by an error probability, which was chosen to be 0.05 in the following.
Ad−Hoc Robust Monolingual English Test Task Top 5 Participants − Standard Recall Levels vs Mean Interpolated Precision
   100%
                                           darmstadt [Experiment DA_4; MAP 45.09%; Not Pooled]
                                           reina [Experiment ROB2; MAP 44.52%; Not Pooled]
    90%                                    uniba [Experiment UNIBAKRF; MAP 42.50%; Not Pooled]
                                           geneva [Experiment ISIENNATTDN; MAP 41.71%; Not Pooled]
                                           know−center [Experiment ASSO; MAP 41.70%; Not Pooled]
    80%


             70%


             60%
 Precision




             50%


             40%


             30%


             20%


             10%


             0%
               0%   10%    20%       30%      40%       50%       60%      70%       80%       90%     100%
                                                       Recall


Fig. 2. Mean average precision of the top 5 participants of the Robust Monolingual
English Task.



We have designed our analysis to follow closely the methodology used by similar
analyses carried out for Text REtrieval Conference (TREC) [23].
    We used the MATLAB Statistics Toolbox, which provides the necessary func-
tionality plus some additional functions and utilities.
    Two tests for goodness of fit to a normal distribution were chosen using the
MATLAB statistical toolbox: the Lilliefors test and the Jarque-Bera test. In the
case of the CLEF tasks under analysis, both tests indicate that the assumption
of normality is not violated for most of the data samples (in this case the runs
for each participant).
    The two tests were:

   – Robust Monolingual vs Robust WSD Monolingual;
   – Robust Bilingual vs Robust WSD Bilingual.

    In both cases, the t-test confirmed that the mean of the two distributions are
different and, in particular, the mean of the monolingual distribution is greater
than the mean of the robust monolingual WSD, and the same happens for the
bilingual. This suggests some loss of performances due to the effect of the word
sense disambiguation in both monolingual and bilingual tasks. However, there
Ad−Hoc Robust WSD Monolingual English Test Task Top 5 Participants − Standard Recall Levels vs Mean Interpolated Precision
     100%
                                 darmstadt [Experiment DA_WSD_4; MAP 45.00%; Not Pooled]
                                 uniba [Experiment UNIBAKEYSYNRF; MAP 43.46%; Not Pooled]
       90%                       know−center [Experiment ASSOWSD; MAP 42.22%; Not Pooled]
                                 reina [Experiment ROBWSD2; MAP 41.23%; Not Pooled]
                                 jaen [Experiment SINAI1_NOWSD_FB_OKAPI; MAP 38.19%; Not Pooled]
       80%


               70%


               60%
   Precision




               50%


               40%


               30%


               20%


               10%


               0%
                 0%   10%     20%      30%       40%      50%       60%      70%       80%      90%      100%
                                                         Recall


Fig. 3. Mean average precision of the top 5 participants of the Robust WSD Monolin-
gual English Task.


are a few topics where the WSD techniques significantly improve the effectiveness
of the retrieval; these are the cases worth studying from a WSD point of view.


3.2            Analysis

In this section we focus on the comparison between WSD and non-WSD runs.
Overall, the best MAP and GMAP results in the monolingual system were for
two distinct runs which did not use WSD information. Several participants were
able to obtain their best MAP and GMAP scores using WSD information. In
the bilingual experiments, the best results in MAP was for non-WSD runs, but
two participants were able to profit from the WSD annotations. As it is difficult
to summarize the behavior of all participants below, we will only mention the
performance of the best teams, as given in Tables 3 and 4. The interested reader
is directed to the working notes of each participant for additional details.
    In the monolingual experiments, cf. Table 3, the best results overall in MAP
was for darmstadt. Their WSD runs scored very similar to the non-WSD runs,
with a slight decrease of MAP (0.09 percentage points) and a slight increase
of GMAP (0.07 percentage points) [15]. The second best MAP score and best
GMAP was attained by reina [16] without WSD, with their WSD systems show-
   Ad−Hoc Robust Bilingual English Test Task Top 5 Participants − Standard Recall Levels vs Mean Interpolated Precision
     100%
                                         reina [Experiment BILI2; MAP 38.42%; Not Pooled]
                                         uniba [Experiment UNIBACROSSKEYRF; MAP 38.09%; Not Pooled]
      90%                                know−center [Experiment BILIASSO; MAP 28.98%; Not Pooled]
                                         ufrgs [Experiment BILINGUAL; MAP 27.65%; Not Pooled]
                                         ixa [Experiment ESENNOWSD; MAP 18.05%; Not Pooled]
      80%


            70%


            60%
Precision




            50%


            40%


            30%


            20%


            10%


            0%
              0%   10%       20%       30%      40%       50%       60%       70%      80%       90%      100%
                                                         Recall


Fig. 4. Mean average precision of the top 5 participants of the Robust Bilingual English
Task.




ing a considerable performance drop. The third best MAP and second GMAP
where obtained by uniba [4] using WSD. This team showed a 0.94 increase in
MAP and 1.67 increase in GMAP with respect to their best non-WSD run. An-
other team showing high MAP and GMAP values was know-center [11], which
attained 0.52 improvements in MAP and 0.83 increase in GMAP with the use
of WSD. Finally, geneva [10] also attained good results, but their WSD system
also had a considerable drop in both MAP and GMAP. All in all, regarding the
use of WSD in the monolingual task, two teams exhibited modest gains, two
teams had quite large performance drops, and the teams reporting best results
had very similar results.
   In the bilingual experiments, cf. Table 4, the best results overall in MAP
were for reina with a system which did not use WSD annotations [16]. The
best GMAP was for geneva using WSD [10]. Unfortunately, they did not submit
any non-WSD run. Uniba [4] got the second best MAP, with better MAP for
the non-WSD run and better GMAP for the WSD run. The differences were
small in both cases (0.56 in MAP, 0.71 in GMAP). Those three teams had the
highest results, well over 35% MAP, and the rest got more modest performances.
know-center [11] reported better results using WSD information (0.66 MAP, 0.26
Ad−Hoc Robust WSD Bilingual English Test Task Top 5 Participants − Standard Recall Levels vs Mean Interpolated Precision
    100%
                        uniba [Experiment UNIBACROSSKEYSYNRF; MAP 37.53%; Not Pooled]
                        geneva [Experiment ISINUSWSDTD; MAP 36.63%; Not Pooled]
     90%                reina [Experiment BILIWSD2; MAP 30.32%; Not Pooled]
                        know−center [Experiment BILIASSOWSD; MAP 29.64%; Not Pooled]
                        ixa [Experiment ESEN1STTOPSBESTSENSE500DOCS; MAP 18.38%; Not Pooled]
     80%


                70%


                60%
    Precision




                50%


                40%


                30%


                20%


                10%


                0%
                  0%   10%   20%      30%      40%       50%       60%      70%       80%       90%     100%
                                                        Recall


Fig. 5. Mean average precision of the top 5 participants of the Robust WSD Bilingual
English Task.


GMAP). Ufrgs [5] only submitted the WSD result. Finally ixa got low results,
with small improvements using WSD information (0.33 MAP, 0.08 GMAP).
   All in all, the exercise showed that some teams did improve results using
WSD (close to 1 MAP point and more than 1 GMAP point in monolingual,
and below 1 MAP/GMAP point in bilingual), but the best results for both
monolingual and bilingual tasks were for systems which did not use WSD.


4               Conclusions
This new edition of the robust WSD exercise has measured to what extent IR
systems could profit from automatic word sense disambiguation information.
The conclusions on the monolingual subtask are similar to the conclusions of
2008. The evidence for using WSD in monolingual IR is mixed, with some top
scoring groups reporting small improvements in MAP and GMAP, but with the
best overall scores for systems not using WSD.
   Regarding the cross-lingual task, the situation is very similar, but the im-
provements reported by using WSD are smaller.
   Instructions and datasets to replicate the results (including Lucene indexes)
are available from http://ixa.si.ehu.es/clirwsd.
5    Acknowledgements
The robust task was partially funded by the Ministry of Education (project
KNOW TIN2006-15049) and the European Commission (project KYOTO ICT-
2007-211423). We want to thank Oier Lopez de Lacalle, who runs the UBC WSD
system, and Yee Seng Chan, Hwee Tou Ng and Zhi Zhong, who run the NUS
WSD system. Their generous contribution was invaluable to run this exercise.


References
 1. Agirre, E., Lopez de Lacalle, O.: UBC-ALM: Combining k-NN with SVD for
    WSD. In Proceedings of the 4th International Workshop on Semantic Evaluations
    (SemEval 2007), Prague, Czech Republic (2007) 341–345
 2. Agirre, E., Magnini, B., Lopez de Lacalle, O., Otegi, A., Rigau, G., Vossen, P.:
    SemEval-2007 Task01: Evaluating WSD on Cross-Language Information Retrieval.
    In Proceedings of CLEF 2007 Workshop, Budapest, Hungary (2007).
 3. Agirre, E., Di Nunzio, G.M., Ferro, N., Peters, C., Mandl, T.: CLEF 2008: Ad Hoc
    Track Overview. In Borri, F., Nardi, A., Peters, C., eds.: Working Notes for the
    CLEF 2009 Workshop, http://www.clef-campaign.org/
 4. Basile, P., Caputo, A., Semeraro, G.: UNIBA-SENSE at CLEF 2009: Robust WSD
    task. In this volume.
 5. Borges, T.B., Moreira, V.P.: UFRGS@CLEF2009: Retrieval by Numbers In this
    volume.
 6. Braschler, M., Peters, C.: CLEF 2003 Methodology and Metrics. In Peters, C.,
    Braschler, M., Gonzalo, J., Kluck, M., eds.: Comparative Evaluation of Multilingual
    Information Access Systems: Fourth Workshop of the Cross–Language Evaluation
    Forum (CLEF 2003) Revised Selected Papers, Lecture Notes in Computer Science
    (LNCS) 3237, Springer, Heidelberg, Germany (2004) 7–20
 7. Chan, Y. S., Ng, H. T., Zhong, Z.: NUS-PT: Exploiting Parallel Texts for Word
    Sense Disambiguation in the English All-Words Tasks. In Proceedings of the 4th
    International Workshop on Semantic Evaluations (SemEval 2007), Prague, Czech
    Republic (2007) 253–256
 8. Cleverdon, C.: The Cranfield Tests on Index Language Devices. In Sparck Jones,
    K., Willett, P., eds.: Readings in Information Retrieval, Morgan Kaufmann Pub-
    lisher, Inc., San Francisco, California, USA (1997) 47–59
 9. Di Nunzio, G.M., Ferro, N.: Appendix C: Results of the Robust Task. In Borri, F.,
    Nardi, A., Peters, C., eds.: Working Notes for the CLEF 2009 Workshop, http:
    //www.clef-campaign.org/ (2008)
10. Guyot, J., Falquet, G., Radhouani, S.: UniGe at CLEF 2009 Robust WSD Task.
    In this volume.
11. Kern, R., Juffinger, A., Granitzer, M.: Application of Axiomatic Approaches to
    Crosslanguage Retrieval. In this volume.
12. Paskin, N., ed.: The DOI Handbook – Edition 4.4.1. International DOI Foundation
    (IDF). http://dx.doi.org/10.1000/186 (2006)
13. Robertson, S.: On GMAP: and Other Transformations. In Yu, P.S., Tsotras, V.,
    Fox, E.A., Liu, C.B., eds.: Proc. 15th International Conference on Information and
    Knowledge Management (CIKM 2006), ACM Press, New York, USA (2006) 78–83
14. Voorhees, E.M.: The TREC Robust Retrieval Track. SIGIR Forum 39 (2005)
    11–20
15. Wolf, E., Bernhard, D., Gurevych, I.: Combining Probabilistic and Translation-
    Based Models for Information Retrieval based on Word Sense Annotations Infor-
    mation Retrieval. In this volume.
16. Zazo, A., Figuerola, C.G., Alonso Berrocal, J.L., Gomez, R.: REINA at CLEF
    2009 Robust-WSD Task: Partial Use of WSD Information for Retrieval. In this
    volume.