=Paper=
{{Paper
|id=Vol-2125/invited_paper_5
|storemode=property
|title=CENTRE@CLEF2018: Overview of the Replicability Task
|pdfUrl=https://ceur-ws.org/Vol-2125/invited_paper_5.pdf
|volume=Vol-2125
|authors=Nicola Ferro,Maria Maistro,Tetsuya Sakai,Ian Soboroff
|dblpUrl=https://dblp.org/rec/conf/clef/FerroMSS18
}}
==CENTRE@CLEF2018: Overview of the Replicability Task==
CENTRE@CLEF2018: Overview of the Replicability Task Nicola Ferro1 , Maria Maistro1 , Tetsuya Sakai2 , and Ian Soboroff3 1 Department of Information Engineering, University of Padua, Italy {ferro,maistro}@dei.unipd.it 2 Waseda University, Japan tetsuyasakai@acm.org 3 National Institute of Standards and Technology (NIST), USA ian.soboroff@nist.gov Abstract. Reproducibility has become increasingly important for many research areas, among those IR is not an exception and has started to be concerned with reproducibility and its impact on research results. This paper describes our first attempt to propose a lab on reproducibility named CENTRE and held during CLEF 2018. The aim of CENTRE is to run a reproducibility challenge across all the major IR evaluation campaigns and to provide the IR community with a venue where previous research results can be explored and discussed. This paper reports the participant results and preliminary considerations on the first edition of CENTRE@CLEF 2018, as well as some suggestions for future editions. 1 Introduction Reproducibility is becoming a primary concern in many areas of science [16] as well as in computer science, as also witnessed by the recent ACM policy on result and artefact review and badging4 . Information Retrieval (IR) is especially interested in reproducibility [12, 13, 34] since it is a discipline strongly rooted in experimentation where experimental evaluation represents a main driver of advancement and innovation. Even if reproducibility has become part of the review forms at major confer- ences like SIGIR, this is more a qualitative assessment performed by a reviewer on the basis of what can be understood from a paper rather than an actual “proof” of the reproducibility of the experiments reported in the paper. Since 2015, the ECIR conference started a new track focused on reproducibility of previously published results. This conference track led to a stable enough flow of 3-4 reproducibility papers accepted each year but, unfortunately, this valu- able effort did not produce a systematic approach to reproducibility: submitting authors adopted different notions of reproducibility, they adopted very diverse experimental protocols, they investigated the most disparate topics, resulting 4 https://www.acm.org/publications/policies/ artifact-review-badging in a very fragmented picture of what was reproduced and what not, and the outcomes of these reproducibility papers are spread over a series of potentially disappearing repositories and Web sites. Moreover, if we consider open source IR systems, they are typically used as: – starting point by new-comers in the field, which take them almost off-the- shelf using default configuration to begin experience with IR and/or specific search tasks; – base system on top of which to add a new component/technique you are interested to develop, keeping all the rest in the default configuration; – baseline for comparison, again using default configuration. Nevertheless, it has been repeatedly shown that best TREC systems still outperform off-the-shelf open source systems [2–4, 23, 24]. This is due to many different factors, among which lack of tuning on a specific collection when using default configuration, but it is also caused by the lack of the specific and advanced components and resources adopted by the best systems. It has been also shown that additivity is an issue, since adding a component on top of a weak or strong base does not produce the same level of gain [4, 23]. This poses a serious challenge when off-the-shelf open source systems are used as stepping stone to test a new component on top of them, because the gain might appear bigger starting from a weak baseline. Overall, the above considerations stress the need and urgency for a systematic approach to reproducibility in IR. Therefore, the goal of CENTRE@CLEF 20185 is to run a joint task across CLEF/NTCIR/TREC on challenging participants: – to reproduce best results of best/most interesting systems in previous edi- tions of CLEF/NTCIR/TREC by using standard open source IR systems; – to contribute back to the community the additional components and re- sources developed to reproduce the results in order to improve existing open source systems. The paper is organized as follows: Section 2 introduces the setup of the lab; Section 3 discusses the participation and the experimental outcomes; and, Section 4 draws some conclusions and outlooks possible future works. 2 Evaluation Lab Setup 2.1 Tasks The CENTRE@CLEF 2018 lab offered two pilot tasks: – Task 1 - Replicability: the task focused on the replicability of selected meth- ods on the same experimental collections; – Task 2 - Reproducibility: the task focused on the reproducibility of selected methods on the different experimental collections. 5 http://www.centre-eval.org/clef2018/ where we adopted the ACM Artifact Review and Badging definition of repli- cability and reproducibility: – Replicability (different team, same experimental setup): the measurement can be obtained with stated precision by a different team using the same mea- surement procedure, the same measuring system, under the same operating conditions, in the same or a different location on multiple trials. For compu- tational experiments, this means that an independent group can obtain the same result using the author’s own artifacts. In CENTRE@CLEF 2018 this meant to use the same collections, topics and ground-truth on which the methods and solutions have been developed and evaluated. – Reproducibility (different team, different experimental setup): The measure- ment can be obtained with stated precision by a different team, a different measuring system, in a different location on multiple trials. For computa- tional experiments, this means that an independent group can obtain the same result using artifacts which they develop completely independently. In CENTRE@CLEF 2018 this meant to use a different experimental col- lection, but in the same domain, from those used to originally develop and evaluate a solution. 2.2 Replicability and Reproducibility Targets For this first edition of CENTRE we prefer to select the target runs among the Ad Hoc tasks in previous editions of CLEF, TREC, and NTCIR. We decided to focus on the Ad Hoc retrieval since it is a general and well known task. Moreover, the algorithms and the approaches used for Ad Hoc retrieval are widely used as basis for all the other types of tasks. For each evaluation campaign, CLEF, TREC, and NTCIR, we considered all the Ad Hoc tracks and we examined the submitted papers and the proposed approaches. CLEF Ad Hoc tracks were proposed from 2000 to 2008, for TREC campaign we focused on the Web Track, from 2009 to 20011, while for NTCIR we checked the WWW track of 2017. To select the target papers among all the submitted ones we considered the following criteria: – the popularity of the track, by accounting for the number of participating groups and the number of submitted runs; – the impact of the proposed approach, measured by the number of citations received by the paper; – the year of publication, since we preferred more recent papers; – the tools used by the author, we discarded all those papers that were not using publicly available retrieval systems as Lucene, Terrier, Solr, and Indri. Below we list the runs selected as targets of replicability and reproducibil- ity among which the participants can choose. For each run, it is specified the collection for replicability and the collections for reproducibility; for more infor- mation, the list also provides references to the papers describing those runs as well as the overviews describing the overall task and collections. Since these runs were not originally thought for being used as targets of a replicability/reproducibility exercise, we contacted the authors of the papers to inform them and ask their consent to use the runs. – Run: AUTOEN [18] • Task type: CLEF Ad Hoc Multilingual Task • Replicability: Multi-8 Two Years On with topics of CLEF 2005 [11] • Reproducibility: Multi-8 with topics of CLEF 2003 [29, 6] – Run: AH-TEL-BILI-X2EN-CLEF2008.TWENTE.FCW [27] • Task type: CLEF Ad Hoc, Bilingual Task • Replicability: TEL English (BL) with topics of CLEF 2008 [1] • Reproducibility: TEL French (BNF) and TEL German (ONB) with topics of CLEF 2008 [1] TEL English (BL), TEL French (BNF) and TEL German (ONB) with topics of CLEF 2009 [15] – Run: AH-TEL-BILI-X2DE-CLEF2008.KARLSRUHE.AIFB ONB EN [30] • Task type: CLEF Ad Hoc, Bilingual Task • Replicability: TEL German (ONB) with topics of CLEF 2008 [1] • Reproducibility: TEL English (BL) and TEL French (BNF) with top- ics of CLEF 2008 [1] TEL English (BL), TEL French (BNF) and TEL German (ONB) with topics of CLEF 2009 [15] – Run: UDInfolabWEB2 [33] • Task type: TREC Ad Hoc Web Task • Replicability: ClueWeb12 Category A with topics of TREC 2013 [9] • Reproducibility: ClueWeb09 Category A and B with topics of TREC 2012 [8] ClueWeb12 Category B with topics of TREC 2013 [9] ClueWeb12 Category A and B with topics of TREC 2014 [10] – Run: uogTrDwl [26] • Task type: TREC Ad Hoc Web Task • Replicability: ClueWeb12 Category A with topics of TREC 2014 [10] • Reproducibility: ClueWeb09 Category A and B with topics of TREC 2012 [8] ClueWeb12 Category A and B with topics of TREC 2013 [9] ClueWeb12 Category B with topics of TREC 2014 [10] – Run: RMIT-E-NU-Own-1 and RMIT-E-NU-Own-3 [17] • Task type: NTCIR Ad Hoc Web Task • Replicability: ClueWeb12 Category B with topics of NTCIR-13 [25] • Reproducibility: ClueWeb12 Category A with topics of NTCIR-13 [25] The participants in CENTRE@CLEF 2018 were provided with the corpora necessary to perform the tasks. In details we made the following collections available: – Multi-8 Two years On, is a document collection containing documents writ- ten in eight languages. A grand total of nearly 1.4 million documents in the languages Dutch, English, Finnish, French, German, Italian, Spanish and Swedish made up the multilingual collection; – TEL data was provided by The European Library, the collection is divided in different subsets, each one corresponding to a different language: English, French and German, each of them containing about one million documents. – ClueWeb09 consists of about 1 billion web pages in ten languages that were collected in January and February 2009. ClueWeb09 Category A represents the whole dataset, while ClueWeb09 Category B consists of the first English segment of Category A, which is roughly the first 50 million pages of the entire dataset; – ClueWeb12 is the successor of ClueWeb2009 and consists of roughly 700 millions English web pages, collected between February 10, 2012 and May 10, 2012. ClueWeb12 Category A represents the whole dataset, while ClueWeb12 Category B is a uniform 7% sample of Category A. Table 1 reports the number of documents and the languages of the documents contained in the provided corpora. Table 1. Corpora used for the first edition of CENTRE@CLEF 2018 with the number of documents and the languages of the documents. Name Year # Documents Languages Multi 8 Two Years On 2005 1,451,643 de, en, es, fi, fr, it, nl, sv TEL English 2008 1,000,100 en TEL French 2008 1,000,100 fr TEL German 2008 869,353 de ClueWeb09, Category A 2010 1,040,809,705 ar, de, en, es, fr, ja, ko, it, pt, zh ClueWeb09, Category B 2010 50,220,423 en ClueWeb12, Category A 2013 733,019,372 en ClueWeb12, Category B 2013 52,343,021 en Finally, Table 2 reports the topics used for the replicability and reproducibil- ity tasks, with the corresponding number of documents, documents’ languages and pool sizes. An example of topic for each evaluation campaign is reported in the Figure 1 for CLEF, 2 for TREC, and 3 for NTCIR. 2.3 Evaluation Measures The quality of the replicability runs has been evaluated from two points of view: Table 2. Topics used for the first edition of CENTRE@CLEF 2018 with the number of documents and the languages of the documents. Evaluation Campaign Task # Topics Languages CLEF 2003 Multi-8 60 en, es CLEF 2005 Multi-8 Two Years On 60 en, es CLEF 2008 TEL bili X2DE 50 en, es, fr CLEF 2008 TEL bili X2EN 153 de, es, fr, CLEF 2008 TEL bili X2FR 50 de, en, es, nl CLEF 2009 TEL bili X2DE 50 en, fr, it, zh CLEF 2009 TEL bili X2EN 153 de, el, fr, it, zh CLEF 2009 TEL bili X2FR 50 de, en, it TREC 2012 Web Track, Ad Hoc Task 50 en TREC 2013 Web Track, Ad Hoc Task 50 en TREC 2014 Web Track, Ad Hoc Task 50 en NTCIR-13 We Want Web Track 100 en456-AH Women's Vote in the USA Find publications on movements or actions aimed at obtaining voting rights for women in the United States Fig. 1. Example of a English topic with its French translation for CLEF 2008, task TEL bili X2DE. 456-AH Le vote des femmes aux États-Unis Trouver des documents sur les mouvements et les actions visant à obtenir le droit de vote pour les femmes aux États-Unis. Fig. 2. Example of topic for TREC 2013, Web Track, Ad Hoc Task. raspberry pi What is a raspberry pi? What is a raspberry pi? What software does a raspberry pi use? What are hardware options for a raspberry pi? How much does a basic raspberry pi cost? Find info about the raspberry pi foundation. Find a picture of a raspberry pi. Fig. 3. Example of topic for NTCIR-13, We Want Web Track. – Effectiveness: how close are the performance scores of the replicated systems to those of the original ones. This is measured using the Root Mean Square Error (RMSE) [22] between the new and original Average Precision (AP) scores: v u m u1 X 2 RMSE = t APorig,i − APreplica,i (1) m i=1 where m is the total number of topics, APorig,i is the AP score of the original target run on topic ti and APreplica,i is the AP score of the replicated run on topic ti . – Ranked result lists: since different result lists may produce the same effective- ness score, we also measure how close are the ranked results list of the repli- cated systems to those of the original ones. This is measured using Kendall’s τ correlation coefficient [21] among the list of retrieved documents for each topic, averaged across all the topics. The Kendall’s τ correlation coefficient on a single topic is given by: P −Q τi orig, replica = q P +Q+T P +Q+U m (2) 1 X τ̄i orig, replica = τi orig, replica m i=1 where m is the total number of topics, P is the total number of concordant pairs (document pairs that are ranked in the same order in both vectors) Q the total number of discordant pairs (document pairs that are ranked in opposite order in the two vectors), T and U are the number of ties, respectively, in the first and in the second ranking. Since for the reproducibility runs we do not have an already existing run to compare against, we planned to compare the reproduced run score with respect to a baseline run to see whether the improvement over the baseline is comparable between the original and the new dataset. However, we did not receive any reproducibility runs so we cannot put in practice this part of the evaluation task. 3 Participation and Outcomes 17 groups registered for participating in CENTRE@CLEF2018 but, unfortu- nately, only one group succeeded in submitting one replicability run. Technical University of Wien (TUW) [19] replicated the run by Cimiano and Sorg, i.e. AH-TEL-BILI-X2DE-CLEF2008.KARLSRUHE.AIFB ONB EN. They submitted four runs described in Table 3, all the code they used to replicate the run is available online6 . 6 https://bitbucket.org/centre_eval/c2018_dataintelligence/src/ master/ Table 3. Submitted run files with their description. Run Name Description esalength100 top10 ESA length: topic k = 1,000 records k = 100, top 10 documents esalength1000 top10 ESA length: topic k = 10,000 records k = 1000, top 10 documents esalength1000 top100 ESA length: topic k = 10,000 records k = 1000, top 100 documents esalength1000 top1000 ESA length: topic k = 10,000 records k = 1000, top 1000 documents The run AH-TEL-BILI-X2DE-CLEF2008.KARLSRUHE.AIFB ONB EN by Cimiano and Sorg uses a cutoff of 1,000 documents and so it has to be compared to esalength1000 top1000, which adopts the same cut-off. However, since TUW submitted runs also for cutoffs 10 and 100 documents, we compare them against versions of the run AH-TEL-BILI-X2DE-CLEF2008.KARLSRUHE.AIFB ONB EN capped at 10 and 100 documents per topic. The paper by Cimiano and Sorg [30] uses Cross-Lingual Explicit Semantic Analysis (CL-ESA) to leverage Wikipedia articles to deal with multiple lan- guages in a uniform way. TUW encountered the following issues in replicating the original run: – the Wikipedia underlying database dump of 2008 was no longer available and they have to resort to the static HTML dump of Wikipedia in the same period; – the above issue caused a processing of Wikipedia articles sensibly different from the original one in [30] and had to rely on several heuristics to cope with HTML; – they fixed an issue in the Inverse Document Frequency (IDF) computation, which might result in negative values according to the equation provided by [30]; – they had to deal with redirect pages in the static HTML dump of Wikipedia in order to find links across wiki pages in multiple languages; – they had to find an alternative interpretation language identification heuris- tics. All these issues prevented TUW from successfully replicating the original run. Indeed the Mean Average Precision (MAP) of the run by Cimiano and Sorg was 0.0667 while the MAP of the run esalength1000 top1000 by TUW is 0.0030. The detailed results at different cutoffs for all the submitted runs are reported in table 4. It clearly emerges that all the above mentioned issues caused the TUW runs to severely underperform with respect to the original run and it is hard to say, in general, the extent to which it is possible to replicate it due to the changes in the language resources available. The difficulties encountered in replicating the run are further stressed by the RMSE between AH-TEL-BILI-X2DE-CLEF2008.KARLSRUHE.AIFB ONB EN and esalength1000 top1000, computed according to eq. (1), which is 0.1132 Table 4. MAP at different cutoff thresholds k = 10, 100, 1000 for the original and the replicated runs. Original Run Replicated Run MAP@10 0.0121 0.0030 esalength100 top10 MAP@10 0.0121 0.0023 esalength1000 top10 MAP@100 0.0465 0.0029 esalength1000 top100 MAP@1000 0.0667 0.0030 esalength1000 top1000 and the average Kendall’s τ correlation among the ranked lists of retrieved doc- uments, computed according to eq. (2), which is −5.69 · 10−04 . Table 5 reports the results of the comparison with RMSE and Kendall’s τ between the original and all the replicated runs. It can be noted how the RMSE deteriorates as the cutoff size increases as it is intuitive since it should be easier to stay closer to the original one when dealing with few top-k documents. Table 5. RMSE and Kendall’s τ between the original and the submitted runs Run Name RMSE Kendall’s τ esalength100 top10 0.0282 −0.0363 esalength1000 top10 0.0283 −0.0474 esalength1000 top100 0.0879 −0.0090 esalength1000 top1000 0.1132 −5.69 · 10−04 Finally, Table 6 shows the Kendall’s τ correlation between the submitted and the original runs computed for each single topic. We can observe as the general trend is to have very low correlations, i.e. very different document rankings between the topics, but there are a few exceptions. For example, topic 467-AH has a correlation of 0.6644 with Explicit Semantic Analysis (ESA) length 1,000 and 10 documents cutoff, which suddenly drops to −0.0253 and −0.0038 for cutoffs 100 and 1,000, respectively, further stressing the fact that it should be easier to replicate the very top-k documents. Another interesting example is topic 490-AH for which the correlation at ESA length 1,000 and 10 documents cutoff is −0.6170, indicating that the right documents have been retrieved but they have been ranked in almost reversed order. 4 Conclusions and Future Work This paper reports the results on the first edition of CENTRE@CLEF2018. A total of 17 participants enrolled in the lab, however just one group managed to submit a run. As reported in the results section, the group encountered many Table 6. Kendall’s τ between the original and the submitted runs computed for each topic. Topic esalength100 top10 esalength1000 top10 esalength1000 top100 esalength1000 top1000 451-AH 0.4377 0.0829 0.0865 -0.0494 452-AH -0.2722 -0.2916 -0.0621 -0.0371 453-AH -0.4639 -0.1030 -0.1718 -0.0197 454-AH 0.2623 -0.3928 -0.0003 0.0310 455-AH -0.0289 0.2416 0.0083 0.0150 456-AH -0.0413 -0.1287 -0.0267 0.0047 457-AH -0.6058 0.3882 0.0703 0.0106 458-AH -0.0227 -0.4998 0.1372 0.0312 459-AH -0.2459 -0.5021 0.0506 0.0337 460-AH 0.2945 -0.0990 -0.0068 0.0137 461-AH -0.1637 -0.5951 -0.1277 0.0134 462-AH -0.0266 -0.0180 0.0922 0.0219 463-AH 0.0827 0.1061 -0.0899 0.0303 464-AH 0.1069 -0.1421 -0.0651 0.0497 465-AH -0.4081 0.1786 0.0257 -0.0182 466-AH -0.3927 -0.2328 -0.0160 0.0153 467-AH 0.2807 0.6644 -0.0253 -0.0038 468-AH -0.0755 -0.0951 -0.0496 0.0040 469-AH -0.3870 -0.3166 0.1479 0.0283 470-AH 0.3339 -0.2227 -0.1398 -0.0076 471-AH 0.2954 0.0493 0.1211 0.0669 472-AH -0.4701 -0.4440 -0.2120 -0.0160 473-AH -0.2483 -0.1736 0.0142 -0.0441 474-AH -0.2340 -0.5220 0.1072 0.0279 475-AH -0.0475 0.2042 -0.0805 0.0316 476-AH 0.1730 -0.0754 0.1722 -0.0044 477-AH -0.5510 -0.0394 0.0025 0.0479 478-AH 0.0264 0.5845 0.1340 0.0453 479-AH 0.4428 -0.0021 -0.0211 -0.0908 480-AH -0.4202 0.0160 0.1304 -0.0682 481-AH -0.4734 0.0248 -0.0674 0.0653 482-AH 0.0462 0.5746 -0.0264 0.0113 483-AH -0.3460 0.0705 -0.0294 -0.0446 484-AH 0.4431 -0.3125 -0.0343 -0.0226 485-AH -0.0411 -0.0088 -0.1511 -0.0502 486-AH 0.1177 -0.1333 -0.1147 -0.0528 487-AH 0.2000 0.2069 -0.2364 -0.0592 488-AH -0.6633 0.3141 0.0783 -0.0360 489-AH 0.8133 0.0726 -0.0114 -0.0076 490-AH -0.0426 -0.6170 0.1077 -0.0296 491-AH -0.7382 -0.1575 -0.0343 -0.0045 492-AH -0.4025 0.1594 -0.0021 0.0290 493-AH 0.1469 0.0091 -0.1079 0.0164 494-AH 0.2364 -0.4308 0.0432 0.0099 495-AH 0.2282 0.4465 0.0442 0.0340 496-AH 0.4478 0.2788 -0.0460 0.0291 497-AH 0.0108 -0.4886 -0.2636 -0.0602 498-AH 0.4813 -0.0073 0.1047 0.0251 499-AH 0.0701 0.1015 0.0759 0.0094 500-AH 0.0212 -0.0954 0.0158 -0.0539 substantial issues which prevented them to actually replicate the targeted run, as described in more detail in their paper [19]. These results support anecdotal evidence in the field about how much diffi- cult is to actually replicate (and even more reproduce) research results, even in a field with such a long experimental tradition as IR is. However, the lack of par- ticipation is a signal that the community is somehow overlooking this important issue. As it also emerged from a recent survey within the SIGIR community [14], while there is a very positive attitude towards reproducibility and it is consid- ered very important from a scientific point of view, there are many obstacles to it such as the effort required to put it into practice, the lack of rewards for achieving it, the possible barriers for new and inexperienced groups, and, least but not last, the (somehow optimistic) researcher’s perception that their own research is already reproducible. For the next edition of the lab we are planning to propose some changes in the lab organization to increase the interest and participation of the research community. First, we will target for newer and more popular systems to be re- produced, moreover we will consider other tasks than the AdHoc, as for example the medical or other popular domains. References 1. Agirre, E., Di Nunzio, G.M., Ferro, N., Mandl, T., Peters, C.: CLEF 2008: Ad Hoc Track Overview. In: Peters, C., Deselaers, T., Ferro, N., Gonzalo, J., Jones, G.J.F., Kurimo, M., Mandl, T., Peñas, A. (eds.) Evaluating Systems for Multilingual and Multimodal Information Access: Ninth Workshop of the Cross–Language Evalua- tion Forum (CLEF 2008). Revised Selected Papers. pp. 15–37. Lecture Notes in Computer Science (LNCS) 5706, Springer, Heidelberg, Germany (2009) 2. Arguello, J., Crane, M., Diaz, F., Lin, J., Trotman, A.: Report on the SIGIR 2015 Workshop on Reproducibility, Inexplicability, and Generalizability of Results (RIGOR). SIGIR Forum 49(2), 107–116 (December 2015) 3. Armstrong, T.G., Moffat, A., Webber, W., Zobel, J.: Has Adhoc Retrieval Im- proved Since 1994? In: Allan, J., Aslam, J.A., Sanderson, M., Zhai, C., Zobel, J. (eds.) Proc. 32nd Annual International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR 2009). pp. 692–693. ACM Press, New York, USA (2009) 4. Armstrong, T.G., Moffat, A., Webber, W., Zobel, J.: Improvements That Don’t Add Up: Ad-Hoc Retrieval Results Since 1998. In: Cheung, D.W.L., Song, I.Y., Chu, W.W., Hu, X., Lin, J.J. (eds.) Proc. 18th International Conference on Infor- mation and Knowledge Management (CIKM 2009). pp. 601–610. ACM Press, New York, USA (2009) 5. Borri, F., Nardi, A., Peters, C., Ferro, N. (eds.): CLEF 2008 Working Notes. CEUR Workshop Proceedings (CEUR-WS.org), ISSN 1613-0073, http://ceur-ws. org/Vol-1174/ (2008) 6. Braschler, M.: CLEF 2003 – Overview of Results. In: Peters et al. [28], pp. 44–63 7. Braschler, M., Di Nunzio, G.M., Ferro, N., Peters, C.: CLEF 2004: Ad Hoc Track Overview and Results Analysis. In: Peters, C., Clough, P., Gonzalo, J., Jones, G.J.F., Kluck, M., Magnini, B. (eds.) Multilingual Information Access for Text, Speech and Images: Fifth Workshop of the Cross–Language Evaluation Forum (CLEF 2004) Revised Selected Papers. pp. 10–26. Lecture Notes in Computer Science (LNCS) 3491, Springer, Heidelberg, Germany (2005) 8. Clarke, C.L.A., Craswell, N., Voorhees, E.M.: Overview of the TREC 2012 Web Track. In: Voorhees, E.M., Buckland, L.P. (eds.) The Twenty-First Text REtrieval Conference Proceedings (TREC 2012). pp. 1–8. National Institute of Standards and Technology (NIST), Special Publication 500-298, Washington, USA (2013) 9. Collins-Thompson, K., Diaz, F., Clarke, C.L.A., Voorhees, E.M.: TREC 2013 Web Track Overview. In: Voorhees [31] 10. Collins-Thompson, K., Macdonald, C., Bennett, P.N., Voorhees, E.M.: TREC 2014 Web Track Overview. In: Voorhees and Ellis [32] 11. Di Nunzio, G.M., Ferro, N., Jones, G.J.F., Peters, C.: CLEF 2005: Ad Hoc Track Overview. In: Peters, C., Gey, F.C., Gonzalo, J., Jones, G.J.F., Kluck, M., Magnini, B., Müller, H., de Rijke, M. (eds.) Accessing Multilingual Information Repositories: Sixth Workshop of the Cross–Language Evaluation Forum (CLEF 2005). Revised Selected Papers. pp. 11–36. Lecture Notes in Computer Science (LNCS) 4022, Springer, Heidelberg, Germany (2006) 12. Ferro, N.: Reproducibility Challenges in Information Retrieval Evaluation. ACM Journal of Data and Information Quality (JDIQ) 8(2), 8:1–8:4 (February 2017) 13. Ferro, N., Fuhr, N., Järvelin, K., Kando, N., Lippold, M., Zobel, J.: Increasing Reproducibility in IR: Findings from the Dagstuhl Seminar on “Reproducibility of Data-Oriented Experiments in e-Science”. SIGIR Forum 50(1), 68–82 (June 2016) 14. Ferro, N., Kelly, D.: SIGIR Initiative to Implement ACM Artifact Review and Badging. SIGIR Forum 52(1) (June 2018) 15. Ferro, N., Peters, C.: CLEF 2009 Ad Hoc Track Overview: TEL & Persian Tasks. In: Peters, C., Di Nunzio, G.M., Kurimo, M., Mandl, T., Mostefa, D., Peñas, A., Roda, G. (eds.) Multilingual Information Access Evaluation Vol. I Text Retrieval Experiments – Tenth Workshop of the Cross–Language Evaluation Forum (CLEF 2009). Revised Selected Papers. pp. 13–35. Lecture Notes in Computer Science (LNCS) 6241, Springer, Heidelberg, Germany (2010) 16. Freire, J., Fuhr, N., Rauber, A. (eds.): Report from Dagstuhl Seminar 16041: Re- producibility of Data-Oriented Experiments in e-Science. Dagstuhl Reports, Vol- ume 6, Number 1, Schloss Dagstuhl–Leibniz-Zentrum für Informatik, Germany (2016) 17. Gallagher, L., Mackenzie, J., Benham, R., Chen, R.C., Scholer, F., Culpepper, J.S.: RMIT at the NTCIR-13 We Want Web Task. In: Kando et al. [20], pp. 402–406 18. Guyot, J., Radhouani, S., Falquet, G.: Ontology-Based Multilingual Informa- tion Retrieval. In: Peters, C., Quochi, V., Ferro, N. (eds.) CLEF 2005 Working Notes. CEUR Workshop Proceedings (CEUR-WS.org), ISSN 1613-0073, http: //ceur-ws.org/Vol-1171/ (2005) 19. Jungwirth, M., Hanbury, A.: Replicating an Experiment in Cross-lingual Informa- tion Retrieval with Explicit Semantic Analysis. In: Cappellato, L., Ferro, N., Nie, J.Y., Soulier, L. (eds.) CLEF 2018 Working Notes. CEUR Workshop Proceedings (CEUR-WS.org), ISSN 1613-0073 (2018) 20. Kando, N., Fujita, S., Kato, M.P., Manabe, T. (eds.): Proc. 13th NTCIR Con- ference on Evaluation of Information Access Technologies. National Institute of Informatics, Tokyo, Japan (2017) 21. Kendall, M.G.: Rank correlation methods. Griffin, Oxford, England (1948) 22. Kenney, J.F., Keeping, E.S.: Mathematics of Statistics – Part One. D. Van Nos- trand Company, Princeton, USA, 3rd edn. (1954) 23. Kharazmi, S., Scholer, F., Vallet, D., Sanderson, M.: Examining Additivity and Weak Baselines. ACM Transactions on Information Systems (TOIS) 34(4), 23:1– 23:18 (June 2016) 24. Lin, J., Crane, M., Trotman, A., Callan, J., Chattopadhyaya, I., Foley, J., Ingersoll, G., Macdonald, C., Vigna, S.: Toward Reproducible Baselines: The Open-Source IR Reproducibility Challenge. In: Ferro, N., Crestani, F., Moens, M.F., Mothe, J., Silvestri, F., Di Nunzio, G.M., Hauff, C., Silvello, G. (eds.) Advances in Informa- tion Retrieval. Proc. 38th European Conference on IR Research (ECIR 2016). pp. 357–368. Lecture Notes in Computer Science (LNCS) 9626, Springer, Heidelberg, Germany (2016) 25. Luo, C., Sakai, T., Liu, Y., Dou, Z., Xiong, C., Xu, J.: Overview of the NTCIR-13 We Want Web Task. In: Kando et al. [20], pp. 394–401 26. McCreadie, R., Deveaud, R., Albakour, D., Mackie, S., Limsopatham, N., Macdon- ald, C., Ounis, I., Thonet, T.: University of Glasgow at TREC 2014: Experiments with Terrier in Contextual Suggestion, Temporal Summarisation and Web Tracks. In: Voorhees and Ellis [32] 27. Nguyen, D., Overwijk, A., Hauff, C., Trieschnigg, D.R., Hiemstra, D., de Jong, F.M.G.: WikiTranslate: Query Translation for Cross-lingual Information Retrieval Using only Wikipedia. In: Borri et al. [5] 28. Peters, C., Braschler, M., Gonzalo, J., Kluck, M. (eds.): Comparative Evaluation of Multilingual Information Access Systems: Fourth Workshop of the Cross–Language Evaluation Forum (CLEF 2003) Revised Selected Papers. Lecture Notes in Com- puter Science (LNCS) 3237, Springer, Heidelberg, Germany (2004) 29. Savoy, J.: Report on CLEF-2003 Multilingual Tracks. In: Peters et al. [28], pp. 64–73 30. Sorg, P., Cimiano, P.: Cross-lingual Information Retrieval with Explicit Semantic Analysis. In: Borri et al. [5] 31. Voorhees, E.M. (ed.): The Twenty-Second Text REtrieval Conference Proceedings (TREC 2013). National Institute of Standards and Technology (NIST), Special Publication 500-302, Washington, USA (2014) 32. Voorhees, E.M., Ellis, A. (eds.): The Twenty-Third Text REtrieval Conference Pro- ceedings (TREC 2014). National Institute of Standards and Technology (NIST), Special Publication 500-308, Washington, USA (2015) 33. Yang, P., Fang, H.: Evaluating the Effectiveness of Axiomatic Approaches in Web Track. In: Voorhees [31] 34. Zobel, J., Webber, W., Sanderson, M., Moffat, A.: Principles for Robust Evaluation Infrastructure. In: Agosti, M., Ferro, N., Thanos, C. (eds.) Proc. Workshop on Data infrastructurEs for Supporting Information Retrieval Evaluation (DESIRE 2011). pp. 3–6. ACM Press, New York, USA (2011) 0008 World Table Tennis Championships