=Paper=
{{Paper
|id=Vol-2696/CLEF2020-preface
|storemode=property
|title=None
|pdfUrl=https://ceur-ws.org/Vol-2696/CLEF2020-preface.pdf
|volume=Vol-2696
|authors=Linda Cappellato,Carsten Eickhoff,Nicola Ferro,Aurélie Névéol
}}
==None==
Preface The CLEF 2020 conference is the twentyfirst edition of the popular CLEF cam- paign and workshop series that has run since 2000 contributing to the systematic evaluation of multilingual and multimodal information access systems, primarily through experimentation on shared tasks. In 2010 CLEF was launched in a new format, as a conference with research presentations, panels, poster and demo sessions and laboratory evaluation workshops. These are proposed and operated by groups of organizers volunteering their time and effort to define, promote, administrate and run an evaluation activity. CLEF 20201 was jointed organized by the Center for Research and Technol- ogy Hellas (CERTH), the University of Amsterdam, and the Democritus Univer- sity of Thrace, and it was expected to be hosted by CERTH, and in particular by the Multimedia Knowledge and Social Media Analytics Laboratory of its In- formation Technologies Institute, at the premises of CERTH, in Thessaloniki, Greece from 22 to 25 September 2020. The outbreak of the Covid-19 pandemic in early 2020 affected the organiza- tion of CLEF 2020. The CLEF steering committee along with the organizers of CLEF 2020, after detailed discussions, decided to run the conference fully virtu- ally. The conference format remained the same as in past years, and consisted of keynotes, contributed papers, lab sessions, and poster sessions, including reports from other benchmarking initiatives from around the world. All sessions were organized and run online. 15 lab proposals were received and evaluated in peer review based on their innovation potential and the quality of the resources created. To identify the best proposals, besides the well-established criteria from the editions of previ- ous years of CLEF such as topical relevance, novelty, potential impact on future world affairs, likely number of participants, and the quality of the organizing consortium, this year we further stressed the connection to real-life usage sce- narios and we tried to avoid as much as possible overlaps among labs in order to promote synergies and integration. The 12 selected labs represented scientific challenges based on new data sets and real world problems in multimodal and multilingual information access. These data sets provide unique opportunities for scientists to explore collections, to develop solutions for these problems, to receive feedback on the performance of their solutions and to discuss the issues with peers at the workshops. We continued the mentorship program to support the preparation of lab proposals for newcomers to CLEF. The CLEF newcomers mentoring program offered help, guidance, and feedback on the writing of draft lab proposals by assigning a mentor to proponents, who helped them in preparing and maturing the lab proposal for submission. If the lab proposal fell into the scope of an 1 http://clef2020.clef-initiative.eu/ Copyright © 2020 for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0). CLEF 2020, 22-25 September 2020, Thessaloniki, Greece. already existing CLEF lab, the mentor helped proponents to get in touch with those lab organizers and team up forces. Building on previous experience, the Labs at CLEF 2020 demonstrate the maturity of the CLEF evaluation environment by creating new tasks, new and larger data sets, new ways of evaluation or more languages. Details of the indi- vidual Labs are described by the Lab organizers in these proceedings. Below is a short summary of them. ARQMath: Answer Retrieval for Mathematical Questions2 considers the problem of finding answers to new mathematical questions among posted answers on the community question answering site Math Stack Exchange. The goals of the lab are to develop methods for mathematical information retrieval based on both text and formula analysis. BioASQ3 challenges researchers with large-scale biomedical semantic index- ing and question answering (QA). The challenges include tasks relevant to hierarchical text classification, machine learning, information retrieval, QA from texts and structured data, multi-document summarization and many other areas. The aim of the BioASQ workshop is to push the research frontier towards systems that use the diverse and voluminous information available online to respond directly to the information needs of biomedical scientists. ChEMU: Information Extraction from Chemical Patents4 proposes two key information extraction tasks over chemical reactions from patents. Task 1 aims to identify chemical compounds and their specific types, i.e. to assign the label of a chemical compound according to the role which it plays within a chemical reaction. Task 2 requires identification of event trigger words (e.g. “added” and “stirred”) which all have the same type of “EVENT TRIGGER”, and then determination of the chemical entity argu- ments of these events. CheckThat!: Identification and Verification of Political Claims5 aims to foster the development of technology capable of both spotting and verify- ing check-worthy claims in political debates in English, Arabic and Italian. The concrete tasks were to assess the check worthiness of a claim in a tweet, check if a (similar) claim has been previously verified, retrieve evidence to fact-check a claim, and verify the factuality of a claim. CLEF eHealth6 aims to support the development of techniques to aid laypeo- ple, clinicians and policy-makers in easily retrieving and making sense of medical content to support their decision making. The goals of the lab are to develop processing methods and resources in a multilingual setting to enrich difficult-to-understand eHealth texts and provide valuable documentation. 2 https://www.cs.rit.edu/~dprl/ARQMath/ 3 http://www.bioasq.org/workshop2020 4 http://chemu.eng.unimelb.edu.au/ 5 https://sites.google.com/view/clef2020-checkthat 6 http://clef-ehealth.org/ eRisk: Early Risk Prediction on the Internet7 explores challenges of evaluation methodology, effectiveness metrics and other processes related to early risk detection. Early detection technologies can be employed in different areas, particularly those related to health and safety. The 2020 edition of the lab focused on texts written in social media for the early detection of signs of self-harm and depression. HIPE: Named Entity Processing on Historical Newspapers8 aims at fostering named entity recognition on heterogeneous, historical and noisy inputs. The goals of the lab are to strengthen the robustness of existing ap- proaches on non-standard input; to enable performance comparison of named entity processing on historical texts; and, in the long run, to foster efficient semantic indexing of historical documents in order to support scholarship on digital cultural heritage collections. ImageCLEF: Multimedia Retrieval9 provides an evaluation forum for vi- sual media analysis, indexing, classification/learning, and retrieval in medi- cal, nature, security and lifelogging applications with a focus on multimodal data, so data from a variety of sources and media. LifeCLEF: Biodiversity Identification and Prediction10 aims at boosting research on the identification and prediction of living organisms in order to solve the taxonomic gap and improve our knowledge of biodiversity. Through its biodiversity informatics related challenges, LifeCLEF is intended to push the boundaries of the state-of-the-art in several research directions at the frontier of multimedia information retrieval, machine learning and knowledge engineering. Lilas: Living Labs for Academic Search11 aims to bring together re- searchers interested in the online evaluation of academic search systems. The long term goal is to foster knowledge on improving the search for aca- demic resources like literature, research data, and the interlinking between these resources in fields from the Life Sciences and the Social Sciences. The immediate goal of this lab is to develop ideas, best practices, and guidelines for a full online evaluation campaign at CLEF 2021. PAN: Digital Text Forensics and Stylometry12 is a networking initiative for the digital text forensics, where researchers and practitioners study tech- nologies that analyze texts with regard to originality, authorship, and trust- worthiness. PAN provides evaluation resources consisting of large-scale cor- pora, performance measures, and web services that allow for meaningful eval- uations. The main goal is to provide for sustainable and reproducible evalu- ations, to get a clear view of the capabilities of state-of-the-art-algorithms. 7 http://erisk.irlab.org/ 8 https://impresso.github.io/CLEF-HIPE-2020/ 9 https://www.imageclef.org/2019 10 http://www.lifeclef.org/ 11 https://clef-lilas.github.io/ 12 http://pan.webis.de/ Touché: Argument retrieval13 is the first shared task on the topic of ar- gument retrieval. Decision making processes, be it at the societal or at the personal level, eventually come to a point where one side will challenge the other with a why-question, which is a prompt to justify one’s stance. Thus, technologies for argument mining and argumentation processing are matur- ing at a rapid pace, giving rise for the first time to argument retrieval. As a group, the 71 lab organizers were based in 14 countries, with Germany, and France leading the distribution. Despite CLEF’s traditionally Europe-based audience, 18 (25.4%) organizers were affiliated with international institutions outside of Europe. The gender distribution was biased towards 81.3% male or- ganizers. CLEF has always been backed by European projects that complement the incredible amount of volunteering work performed by Lab Organizers and the CLEF community with the resources needed for its necessary central coordina- tion, in a similar manner to the other major international evaluation initiatives such as TREC, NTCIR, FIRE and MediaEval. Since 2014, the organisation of CLEF no longer has direct support from European projects and are working to transform itself into a self-sustainable activity. This is being made possible thanks to the establishment of the CLEF Association14 , a non-profit legal entity in late 2013, which, through the support of its members, ensures the resources needed to smoothly run and coordinate CLEF. Acknowledgments We would like to thank the mentors who helped in shepherding the preparation of lab proposals by newcomers: Lorraine Goeuriot, Université Grenoble Alpes, France Frank Hopfgartner, Uni- versity of Sheffield, UK; Jaap Kamps, University of Amsterdam, The Netherlands; Josiane Mothe, IRIT, Université de Toulouse, France; Henning Müller, University of Applied Sciences Western Switzerland (HES-SO), Switzerland. We would like to thank the members of CLEF-LOC (the CLEF Lab Organi- zation Committee) for their thoughtful and elaborate contributions to assessing the proposals during the selection process: Marianna Apidianaki, University of Helsinki, Finland; Martin Braschler, Zurich University of Applied Sciences, Switzerland; Ingo Frommholz, University of Bedfordshire, UK; Donna Harman, National Institute of Standards and Technology (NIST), USA; Morgan Harvey, University of Sheffield, UK; Jiyin He, Signal AI, UK 13 https://events.webis.de/touche-20/ 14 http://www.clef-initiative.eu/association Rezarta Islamaj Doğan, National Library of Medicine, USA; Yue Ma, University Paris Sud, France; Henning Müller, University of Applied Sciences Western Switzerland (HES-SO), Switzerland; Maarten de Rijke, University of Amsterdam, The Netherlands. Last but not least, without the important and tireless effort of the enthu- siastic and creative proposal authors, the organizers of the selected labs and workshops, the colleagues and friends involved in running them, and the partic- ipants who contribute their time to making the labs and workshops a success, the CLEF labs would not be possible. Thank you all very much! September, 2020 Linda Cappellato Carsten Eickhoff Nicola Ferro Aurélie Névéol Organization CLEF 2020, Conference and Labs of the Evaluation Forum – Experimental IR meets Multilinguality, Multimodality, and Interaction, was hosted (online) by the Multimedia Knowledge and Social Media Analytics Laboratory (MKLab) of the Information Technologies Institute (ITI) of the Center for Research and Technology Hellas (CERTH), Thessaloniki, Greece. General Chairs Evangelos Kanoulas, Univ. of Amsterdam, the Netherlands Theodora Tsikrika, Information Technologies Institute, CERTH, Greece Stefanos Vrochidis, Information Technologies Institute, CERTH, Greece Avi Arampatzis, Democritus University of Thrace, Greece Program Chairs Hideo Joho, University of Tsukuba, Japan Christina Lioma, University of Copenhagen, Denmark Lab Chairs Aurélie Névéol, Université Paris Saclay, CNRS, LIMSI, France Carsten Eickhoff, Brown University, USA Lab Mentorship Chair Lorraine Goeuriot, Université Grenoble Alpes, France Proceedings Chairs Linda Cappellato, University of Padua, Italy Nicola Ferro, University of Padua, Italy Local Organization Vivi Ntrigkogia, Information Technologies Institute, CERTH, Greece CLEF Steering Committee Steering Committee Chair Nicola Ferro, University of Padua, Italy Deputy Steering Committee Chair for the Conference Paolo Rosso, Universitat Politècnica de València, Spain Deputy Steering Committee Chair for the Evaluation Labs Martin Braschler, Zurich University of Applied Sciences, Switzerland Members Khalid Choukri, Evaluations and Language resources Distribution Agency (ELDA), France Paul Clough, University of Sheffield, United Kingdom Fabio Crestani, Università della Svizzera italiana, Switzerland Norbert Fuhr, University of Duisburg-Essen, Germany Lorraine Goeuriot, Université Grenoble Alpes, France Julio Gonzalo, National Distance Education University (UNED), Spain Donna Harman, National Institute for Standards and Technology (NIST), USA Djoerd Hiemstra, University of Twente, The Netherlands Evangelos Kanoulas, University of Amsterdam, The Netherlands Birger Larsen, University of Aalborg, Denmark David E. Losada, Universidade de Santiago de Compostela, Spain Mihai Lupu, Vienna University of Technology, Austria Josiane Mothe, IRIT, Université de Toulouse, France Henning Müller, University of Applied Sciences Western Switzerland (HES-SO), Switzerland Jian-Yun Nie, Université de Montréal, Canada Eric SanJuan, University of Avignon, France Giuseppe Santucci, Sapienza University of Rome, Italy Jacques Savoy, University of Neuchêtel, Switzerland Laure Soulier, Pierre and Marie Curie University (Paris 6), France Christa Womser-Hacker, University of Hildesheim, Germany Past Members Jaana Kekäläinen, University of Tampere, Finland Séamus Lawless, Trinity College Dublin, Ireland Carol Peters, ISTI, National Council of Research (CNR), Italy (Steering Committee Chair 2000–2009) Emanuele Pianta, Centre for the Evaluation of Language and Communication Technologies (CELCT), Italy Maarten de Rijke, University of Amsterdam UvA, The Netherlands Alan Smeaton, Dublin City University, Ireland