=Paper= {{Paper |id=Vol-1180/CLEF2014wn-All-CappellatoEt2014 |storemode=property |title=Preface |pdfUrl=https://ceur-ws.org/Vol-1180/CLEF2014wn-All-CappellatoEt2014.pdf |volume=Vol-1180 |dblpUrl=https://dblp.org/rec/conf/clef/CappellatoFHK14 }} ==Preface== https://ceur-ws.org/Vol-1180/CLEF2014wn-All-CappellatoEt2014.pdf
                                      Preface



The CLEF 2014 conference is the fifteenth edition of the popular CLEF campaign and
workshop series which has run since 2000 contributing to the systematic evaluation of
multilingual and multimodal information access systems, primarily through experi-
mentation on shared tasks. In 2010 CLEF was launched in a new format, as a confer-
ence with research presentations, panels, poster and demo sessions and laboratory
evaluation workshops. These are proposed and operated by groups of organizers vol-
unteering their time and effort to define, promote, administrate and run an evaluation
activity.

Out of 10 received lab proposals, 8 laboratories have been selected and run during
CLEF 2014. To identify the best proposals, besides well-established criteria from
previous years' editions of CLEF such as topical relevance, novelty, potential impact
on future world affairs, likely number of participants, and the quality of the organiz-
ing consortium. This year we further stressed the connection to real-life usage scenar-
ios and we tried to avoid as much as possible overlaps among labs in order to promote
synergies and integration. This was possible also thanks to a new activity introduced
in 2013, which is a lab organizers and proposers meeting co-located with the Europe-
an Conference on Information Retrieval (ECIR), held in Moscow on March 26th 2013
and supported by the ELIAS network1.

CLEF has been always backed by European projects which complemented the incred-
ible amount of volunteering work performed by Lab Organizers and the CLEF com-
munity with the resources needed for its necessary central coordination, as it happens
for the other major evaluation initiatives as TREC, NTCIR and FIRE. Since 2014,
CLEF no longer has support from European projects and it has made an effort to turn
itself into a self-sustainable activity. This has been possible thanks to the establish-
ment in late 2013 of the CLEF Association, a no profit legal entity, which, through
the support of its members, ensures the resources needed to smoothly run and coordi-
nate CLEF.

Each Lab, building on previous experience, demonstrated maturity coming with new
tasks, new and larger data sets, new ways of evaluation or more languages. They are
described by the Lab organizers in details, here we just brief on them.




1
    http://www.elias-network.eu/
CLEF eHealth - ShARe/CLEF eHealth Evaluation Lab: The usage scenario of the
CLEF eHealth lab is to ease patients and next-of-kins’ ease in understanding eHealth
information. The lab contains three tasks: Visual-Interactive Search and Exploration
of eHealth Data, Information extraction from clinical text, User-centered health in-
formation retrieval.


ImageCLEF: ImageCLEF aims at providing benchmarks for the challenging task of
image annotation for a wide range of source images and annotation objective. The
tasks address different aspects of the annotation problem and are aimed at supporting
and promoting the cutting-edge research addressing the key challenges in the field,
such as multi-modal image annotation, domain adaptation and ontology driven image
annotation. The Lab tasks are: Robot Vision, Scalable concept Image Annotation,
Liver CT Annotation and Domain Adaptation.


INEX — Initiative for the Evaluation of XML retrieval: INEX builds evaluation
benchmarks for search in the context of rich structure such as document structure,
semantic metadata, entities, or genre/topical structure. INEX 2014 runs the following
tasks: Social Book Search Task, Interactive Social Book Search Task, Tweet Contex-
tualization Task.


LifeCLEF: LifeCLEF aims at evaluating multimedia analysis and retrieval tech-
niques on biodiversity data for species identification. The three Tasks regard respec-
tively birds, plants and fishes: BirdCLEF (a bird songs identification task based on
Xeno-Canto audio recordings), PlantCLEF (an image-based plant identification task
based on the data of Tela Botanica social network) and FishCLEF (a fish video sur-
veillance task based on the data of the Fish4Knowledge network).


NEWSREEL - News Recommendation Evaluation Lab: Newsreel Lab is the first
news recommendation evaluation Lab and it is divided in two different tasks. One
task is about predicting the items a user will click in the next 10 Minutes based on the
offline dataset. The other task is about predicting the articles users will click; in this
task, the prediction algorithms are evaluated in an online scenario based on live user-
interactions.


PAN Lab on Uncovering Plagiarism, Authorship, and Social Software Misuse:
PAN centers around the topics of plagiarism, authorship, and social software misuse.
The goal is to foster research on automatic detection and uncovering. People increas-
ingly share their work online, contribute to open projects and engage in web-based
social interactions. The ease and anonymity with which this can be done raises con-
cerns about verifiability and trust: Is a given text an original? Is the author the one
who she claims to be? Does a piece of information come from a trusted source? There
are three Lab Tasks: Author Identification, Author Profiling and Plagiarism Detec-
tion.


QA Track — CLEF Question Answering Track: In the current general scenario for
the CLEF QA Track, the starting point is always a Natural Language question. How-
ever, answering some questions may need to query Linked Data; whereas some ques-
tions may need textual inferences and querying free-text. Answering some queries
may need both. The tasks are: QALD - Question Answering over Linked Data, Bio-
ASQ - Biomedical semantic indexing and question answering, Entrance Exams.


RepLab: The aim of RepLab is to bring together the Information Access research
community with representatives from the Online Reputation Management industry,
with the ultimate goals of (i) establishing a roadmap on the topic that includes a de-
scription of the language technologies required in terms of resources, algorithms, and
applications; (ii) specifying suitable evaluation methodologies and metrics to measure
scientific progress; and (iii) developing of test collections that enable systematic com-
parison of algorithms and reliable benchmarking of commercial systems. Lab Tasks
are Reputation Dimensions and Author Profiling.
Acknowledgements

We would like to thank the members of CLEF-LOC (the CLEF Lab Organization
Committee) for their thoughtful and elaborate contributions to assessing the proposals
during the selection process:

Donna Harman, National Institute for Standards and Technology (NIST), USA

Carol Peters, ISTI, National Council of Research (CNR), Italy

Marteen de Rijke, University of Amsterdam, The Netherlands

Last but not least without the important and tireless effort of the enthusiastic and crea-
tive proposal authors, the organizers of the selected labs, the colleagues and friends
involved in running them, and the participants who contribute their time to making
the labs and workshops a success, the CLEF labs would not be possible.

Thank you all very much!




July 2014



                                                       Martin Halvey and Wessel Kraaij
                                                                          Lab Chairs

                                                     Linda Cappellato and Nicola Ferro
                                                               Working Notes Editors