Preface The CLEF 2017 conference is the eighteenth edition of the popular CLEF cam- paign and workshop series which has run since 2000 contributing to the sys- tematic evaluation of multilingual and multimodal information access systems, primarily through experimentation on shared tasks. In 2010 CLEF was launched in a new format, as a conference with research presentations, panels, poster and demo sessions and laboratory evaluation workshops. These are proposed and operated by groups of organizers volunteering their time and effort to define, promote, administrate and run an evaluation activity. CLEF 20171 was hosted by the ADAPT Centre2 , Dublin City University and Trinity College Dublin from the 11th to 14th September 2017. This year’s conference was also co-located with MediaEval3 and the program included joint sessions between both MediaEval and CLEF to allow for cross fertilisation. Eight laboratories and two workshops were selected and run during CLEF 2017. To identify the best proposals, besides well-established criteria from previ- ous years’ editions of CLEF such as topical relevance, novelty, potential impact on future world affairs, likely number of participants, and the quality of the organizing consortium. This year we further stressed the connection to real-life usage scenarios and we tried to avoid as much as possible overlaps among labs in order to promote synergies and integration. The Labs at CLEF 2017, building on previous experience, demonstrate the maturity of the CLEF evaluation environment via the incorporation of new tasks, new and larger data sets, new ways of evaluation or more languages. Details of the individual Labs are described by the Lab organizers in these proceedings, here we just provide brief comment on each one. News Recommendation Evaluation Lab (NEWSREEL)4 provides a vehicle for the IR/recommender system communities to move from conventional offline evaluation to online evaluation. We address the following information access challenge: Whenever a visitor of an online news portal reads a news article on their side, the task is to recommend other news articles that the user might be interested in. LifeCLEF5 aims at boosting research on the identification of living organ- isms and on the production of biodiversity data in general. Through its biodiver- sity informatics related challenges, LifeCLEF is intended to push the boundaries 1 http://clef2017.clef-initiative.eu/ 2 http://adaptcentre.ie/ 3 http://www.multimediaeval.org/ 4 http://clef-newsreel.org/ 5 http://www.lifeclef.org/ of the state-of-the-art in several research directions at the frontier of multimedia information retrieval, machine learning and knowledge engineering. Uncovering Plagiarism, Authorship and Social Software Misuse (PAN)6 provides evaluation of uncovering plagiarism, authorship, and social software misuse. PAN offers three tasks at CLEF 2017 with new evaluation re- sources consisting of large-scale corpora, performance measures, and web services that allow for meaningful evaluations. The main goal is to provide for sustainable and reproducible evaluations, to get a clear view of the capabilities of state-of- the-art-algorithms. The tasks are: author identification; author profiling; and, author obfuscation. CLEFeHealth7 provides scenarios which aim to ease patients and nurses understanding and accessing of eHealth information. The goals of the lab are to develop processing methods and resources in a multilingual setting to enrich difficult-to-understand eHealth texts, and provide valuable documentation. The tasks are: multilingual Information extraction; technologically assisted reviews in empirical medicine; and, patient-centred information retrieval. Cultural Microblog Contextualization (CMC) Workshop8 deals with how cultural context of a microblog affects its social impact at large. This in- volves microblog search, classification, filtering, language recognition, localiza- tion, entity extraction, linking open data and summarization. Regular Lab par- ticipants have access to the private massive multilingual microblog stream of The festival galleries project. ImageCLEF9 organises 3 main tasks with a global objective of benchmark- ing lifelogging retrieval and summarization, tuberculosis type prediction from CT images and bio-medical image caption prediction; and a pilot task on re- mote sensing image analysis. Early risk prediction on the Internet (eRisk)10 explores issues of eval- uation methodology, effectiveness metrics and other processes related to early risk detection. Early detection technologies can be employed in different areas, particularly those related to health and safety. For instance, early alerts could be sent when a predator starts interacting with a child for sexual purposes, or when a potential offender starts publishing antisocial threats on a blog, forum or social network. Our main goal is to pioneer a new interdisciplinary research area that would be potentially applicable to a wide variety of situations and to many different personal profiles. Personalised Information Retrieval at CLEF (PIR-CLEF)11 provides a framework for evaluation of Personalised Information Retrieval (PIR). Current 6 http://pan.webis.de/ 7 https://sites.google.com/site/clefehealth2017/ 8 https://mc2.talne.eu/ 9 http://imageclef.org/2017 10 http://early.irlab.org/ 11 http://www.ir.disco.unimib.it/pirclef2017/ approaches to the evaluation of PIR are user-centered, i.e., they rely on exper- iments that involve real users in a supervised environment, PiR-CLEF aims to develop and demonstrate a methodology for evaluation PIR which enables re- peatable experiments to enable the detailed exploration of personal models and their exploitation in IR. Dynamic Search for Complex Tasks12 Information Retrieval research has traditionally focused on serving the best results for a single query – so- called ad hoc retrieval. However, users typically search iteratively, refining and reformulating their queries during a session. A key challenge in the study of this interaction is the creation of suitable evaluation resources to assess the effectiveness of IR systems over sessions. The goal of the CLEF Dynamic Search lab is to propose and standardize an evaluation methodology that can lead to reusable resources and evaluation metrics able to assess retrieval performance over an entire session, keeping the “user” in the loop. Multimodal Spatial Role Labeling13 explores the extraction of spatial information from two information resources that is image and text. This is im- portant for various applications such as semantic search, question answering, geographical information systems and even in robotics for machine understand- ing of navigational instructions or instructions for grabbing and manipulating objects. CLEF has been always backed by European projects which complement the incredible amount of volunteering work performed by Lab Organizers and the CLEF community with the resources needed for its necessary central coordina- tion, in a similar manner to the other major international evaluation initiatives as TREC, NTCIR, FIRE and MediaEval. Since 2014, the organisation of CLEF no longer has direct support from European projects and working to transform itself into a self-sustainable activity. This is being made possible thanks to the establishment in late 2013 of the CLEF Association14 , a non-profit legal entity, which, through the support of its members, ensures the resources needed to smoothly run and coordinate CLEF. Acknowledgments We would like to thank the members of CLEF-LOC (the CLEF Lab Organiza- tion Committee) for their thoughtful and elaborate contributions to assessing the proposals during the selection process: Martin Braschler, Zurich University of Applied Sciences, Switzerland Donna Harman, National Institute for Standards and Technology (NIST), USA Marteen de Rijke, University of Amsterdam, The Netherlands 12 https://ekanou.github.io/dynamicsearch/ 13 http://www.cs.tulane.edu/~pkordjam/mSpRL_CLEF_lab.htm/ 14 http://www.clef-initiative.eu/association Last but not least without the important and tireless effort of the enthusiastic and creative proposal authors, the organizers of the selected labs and workshops, the colleagues and friends involved in running them, and the participants who contribute their time to making the labs and workshops a success, the CLEF labs would not be possible. Thank you all very much! July, 2017 Linda Cappellato Nicola Ferro Lorraine Goeuriot Thomas Mandl Organization CLEF 2017, Conference and Labs of the Evaluation Forum – Experimental IR meets Multilinguality, Multimodality, and Interaction, was organized by the ADAPT Centre in Dublin City University and Trinity College, Dublin, Ireland. General Chair Gareth J. F. Jones, Dublin City University, Ireland Séamus Lawless, Trinity College Dublin, Ireland Program Chair Julio Gonzalo, UNED, Spain Liadh Kelly, Dublin City University, Ireland Lab Chair Lorraine Goeuriot, Université Grenoble Alpes, France Thomas Mandl, University of Hildesheim, Germany Proceedings Chairs Linda Cappellato, University of Padua, Italy Nicola Ferro, University of Padua, Italy Local Organization Piyush Arora, Dublin City University, Ireland Mostafa Bayomi, Trinity College Dublin, Ireland Annalina Caputo, Trinity College Dublin, Ireland Joris Vreeke (webmaster), ADAPT Centre, Ireland CLEF Steering Committee Steering Committee Chair Nicola Ferro, University of Padua, Italy Deputy Steering Committee Chair for the Conference Paolo Rosso, Universitat Politècnica de València, Spain Deputy Steering Committee Chair for the Evaluation Labs Martin Braschler, Zurich University of Applied Sciences, Switzerland Members Khalid Choukri, Evaluations and Language resources Distribution Agency (ELDA), France Paul Clough, University of Sheffield, United Kingdom Norbert Fuhr, University of Duisburg-Essen, Germany Julio Gonzalo, National Distance Education University (UNED), Spain Donna Harman, National Institute for Standards and Technology (NIST), USA Djoerd Hiemstra, University of Twente, The Netherlands Evangelos Kanoulas, University of Amsterdam, The Netherlands Birger Larsen, University of Aalborg, Denmark Mihai Lupu, Vienna University of Technology, Austria Josiane Mothe, IRIT, Université de Toulouse, France Henning Müller, University of Applied Sciences Western Switzerland (HES-SO), Switzerland Maarten de Rijke, University of Amsterdam UvA, The Netherlands Giuseppe Santucci, Sapienza University of Rome, Italy Jacques Savoy, University of Neuchêtel, Switzerland Christa Womser-Hacker, University of Hildesheim, Germany Past Members Jaana Kekäläinen, University of Tampere, Finland Carol Peters, ISTI, National Council of Research (CNR), Italy (Steering Committee Chair 2000–2009) Emanuele Pianta, Centre for the Evaluation of Language and Communication Technologies (CELCT), Italy Alan Smeaton, Dublin City University, Ireland