<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Extended overview of the CLEF 2024 LongEval Lab on Longitudinal Evaluation of Model Performance Notebook for the LongEval Lab at CLEF 2024</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Rabab</forename><surname>Alkhalifa</surname></persName>
							<email>raalkhalifa@iau.edu.sa</email>
							<affiliation key="aff0">
								<orgName type="institution">Queen Mary University of London</orgName>
								<address>
									<country key="GB">UK</country>
								</address>
							</affiliation>
							<affiliation key="aff1">
								<orgName type="institution">Imam Abdulrahman Bin Faisal University</orgName>
								<address>
									<country>SA</country>
								</address>
							</affiliation>
							<affiliation key="aff14">
								<orgName type="department">Institute of Engineering</orgName>
								<orgName type="institution">Univ. Grenoble Alpes</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Hsuvas</forename><surname>Borkakoty</surname></persName>
							<email>borkakotyh@cardiff.ac.uk</email>
							<affiliation key="aff2">
								<orgName type="institution">Cardiff University</orgName>
								<address>
									<country key="GB">UK</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Romain</forename><surname>Deveaud</surname></persName>
							<email>r.deveaud@qwant.com</email>
							<affiliation key="aff3">
								<address>
									<settlement>Qwant</settlement>
									<country key="FR">France</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Alaa</forename><surname>El-Ebshihy</surname></persName>
							<affiliation key="aff4">
								<orgName type="institution" key="instit1">Research Studios Austria</orgName>
								<orgName type="institution" key="instit2">Data Science Studio</orgName>
								<address>
									<settlement>Vienna</settlement>
									<region>AT</region>
								</address>
							</affiliation>
							<affiliation key="aff5">
								<orgName type="institution">TU Wien</orgName>
								<address>
									<country key="AT">Austria</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Luis</forename><surname>Espinosa-Anke</surname></persName>
							<email>espinosa-ankel@cardiff.ac.uk</email>
							<affiliation key="aff2">
								<orgName type="institution">Cardiff University</orgName>
								<address>
									<country key="GB">UK</country>
								</address>
							</affiliation>
							<affiliation key="aff6">
								<orgName type="laboratory">AMPLYFI</orgName>
								<address>
									<country key="GB">UK</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Tobias</forename><surname>Fink</surname></persName>
							<email>tobias.fink@researchstudio.at</email>
							<affiliation key="aff4">
								<orgName type="institution" key="instit1">Research Studios Austria</orgName>
								<orgName type="institution" key="instit2">Data Science Studio</orgName>
								<address>
									<settlement>Vienna</settlement>
									<region>AT</region>
								</address>
							</affiliation>
							<affiliation key="aff5">
								<orgName type="institution">TU Wien</orgName>
								<address>
									<country key="AT">Austria</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Petra</forename><surname>Galuščáková</surname></persName>
							<email>petra.galuscakova@uis.no</email>
							<affiliation key="aff9">
								<orgName type="institution">University of Stavanger</orgName>
								<address>
									<settlement>Stavanger</settlement>
									<country key="NO">Norway</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Gabriela</forename><surname>Gonzalez-Saez</surname></persName>
							<email>gabriela-nicole.gonzalez-saez@univ-grenoble-alpes.fr</email>
							<affiliation key="aff7">
								<orgName type="department">INP 1</orgName>
								<orgName type="institution" key="instit1">Univ. Grenoble Alpes</orgName>
								<orgName type="institution" key="instit2">CNRS</orgName>
								<address>
									<settlement>Grenoble</settlement>
								</address>
							</affiliation>
							<affiliation key="aff8">
								<orgName type="institution">LIG</orgName>
								<address>
									<settlement>Grenoble</settlement>
									<country key="FR">France</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Lorraine</forename><surname>Goeuriot</surname></persName>
							<email>lorraine.goeuriot@univ-grenoble-alpes.fr</email>
							<affiliation key="aff7">
								<orgName type="department">INP 1</orgName>
								<orgName type="institution" key="instit1">Univ. Grenoble Alpes</orgName>
								<orgName type="institution" key="instit2">CNRS</orgName>
								<address>
									<settlement>Grenoble</settlement>
								</address>
							</affiliation>
							<affiliation key="aff8">
								<orgName type="institution">LIG</orgName>
								<address>
									<settlement>Grenoble</settlement>
									<country key="FR">France</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">David</forename><surname>Iommi</surname></persName>
							<email>david.iommi@researchstudio.at</email>
							<affiliation key="aff4">
								<orgName type="institution" key="instit1">Research Studios Austria</orgName>
								<orgName type="institution" key="instit2">Data Science Studio</orgName>
								<address>
									<settlement>Vienna</settlement>
									<region>AT</region>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Maria</forename><surname>Liakata</surname></persName>
							<email>m.liakata@qmul.ac.uk</email>
							<affiliation key="aff0">
								<orgName type="institution">Queen Mary University of London</orgName>
								<address>
									<country key="GB">UK</country>
								</address>
							</affiliation>
							<affiliation key="aff10">
								<orgName type="institution">Alan Turing Institute</orgName>
								<address>
									<country key="GB">UK</country>
								</address>
							</affiliation>
							<affiliation key="aff11">
								<orgName type="institution">University of Warwick</orgName>
								<address>
									<country key="GB">UK</country>
								</address>
							</affiliation>
							<affiliation key="aff14">
								<orgName type="department">Institute of Engineering</orgName>
								<orgName type="institution">Univ. Grenoble Alpes</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Harish</forename><forename type="middle">Tayyar</forename><surname>Madabushi</surname></persName>
							<affiliation key="aff12">
								<orgName type="institution">University of Bath</orgName>
								<address>
									<country key="GB">UK</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Pablo</forename><surname>Medina-Alias</surname></persName>
							<affiliation key="aff12">
								<orgName type="institution">University of Bath</orgName>
								<address>
									<country key="GB">UK</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Philippe</forename><surname>Mulhem</surname></persName>
							<email>philippe.mulhem@imag.fr</email>
							<affiliation key="aff7">
								<orgName type="department">INP 1</orgName>
								<orgName type="institution" key="instit1">Univ. Grenoble Alpes</orgName>
								<orgName type="institution" key="instit2">CNRS</orgName>
								<address>
									<settlement>Grenoble</settlement>
								</address>
							</affiliation>
							<affiliation key="aff8">
								<orgName type="institution">LIG</orgName>
								<address>
									<settlement>Grenoble</settlement>
									<country key="FR">France</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Florina</forename><surname>Piroi</surname></persName>
							<email>florina.piroi@researchstudio.at</email>
							<affiliation key="aff4">
								<orgName type="institution" key="instit1">Research Studios Austria</orgName>
								<orgName type="institution" key="instit2">Data Science Studio</orgName>
								<address>
									<settlement>Vienna</settlement>
									<region>AT</region>
								</address>
							</affiliation>
							<affiliation key="aff5">
								<orgName type="institution">TU Wien</orgName>
								<address>
									<country key="AT">Austria</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Martin</forename><surname>Popel</surname></persName>
							<email>popel@ufal.mff.cuni.cz</email>
							<affiliation key="aff13">
								<orgName type="institution">Charles University</orgName>
								<address>
									<settlement>Prague</settlement>
									<country key="CZ">Czech Republic</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Arkaitz</forename><surname>Zubiaga</surname></persName>
							<email>a.zubiaga@qmul.ac.uk</email>
							<affiliation key="aff0">
								<orgName type="institution">Queen Mary University of London</orgName>
								<address>
									<country key="GB">UK</country>
								</address>
							</affiliation>
							<affiliation key="aff14">
								<orgName type="department">Institute of Engineering</orgName>
								<orgName type="institution">Univ. Grenoble Alpes</orgName>
							</affiliation>
						</author>
						<title level="a" type="main">Extended overview of the CLEF 2024 LongEval Lab on Longitudinal Evaluation of Model Performance Notebook for the LongEval Lab at CLEF 2024</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">A8CECFA4649940763780364EC99803A1</idno>
					<idno type="DOI">10.48436/xr350-79683</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T17:59+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Evaluation, Temporal Persistence, Temporal Generalisability, Information Retrieval, Text Classification A. Zubiaga) 0000-0002-2875-5400 (R. Alkhalifa)</term>
					<term>0000-0003-3262-0127 (H. Borkakoty)</term>
					<term>0000-0003-2676-7405 (R. Deveaud)</term>
					<term>0000-0001-6644-2360 (A. El-Ebshihy)</term>
					<term>0000-0001-6830-9176 (L. Espinosa-Anke)</term>
					<term>0000-0002-1045-8352 (T. Fink)</term>
					<term>0000-0001-6328-7131 (P. Galuščáková)</term>
					<term>0000-0003-0878-5263 (G. Gonzalez-Saez)</term>
					<term>0000-0001-7491-1980 (L. Goeuriot)</term>
					<term>0000-0002-4270-5709 (D. Iommi)</term>
					<term>0000-0001-5765-0416 (M. Liakata)</term>
					<term>0000-0001-5260-3653 (H. T. Madabushi)</term>
					<term>0009-0001-4202-8664 (P. Medina-Alias)</term>
					<term>0000-0002-3245-6462 (P. Mulhem)</term>
					<term>0000-0001-7584-6439 (F. Piroi)</term>
					<term>0000-0002-3628-8419 (M. Popel)</term>
					<term>0000-0003-4583-3623 (A. Zubiaga)</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>We describe the second edition of the LongEval CLEF 2024 shared task. This lab evaluates the temporal persistence of Information Retrieval (IR) systems and Text Classifiers. Task 1 requires IR systems to run on corpora acquired at several timestamps, and evaluates the drop in system quality (NDCG) along these timestamps. Task 2 tackles binary sentiment classification at different points in time, and evaluates the performance drop for different temporal gaps. Overall, 37 teams registered for Task 1 and 25 for Task 2. Ultimately, 14 and 4 teams participated in Task 1 and Task 2, respectively.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Outside the strict scientific context, the European Artificial Intelligence Act<ref type="foot" target="#foot_0">1</ref> , adopted by European Commission in 2024, stresses in Article 17, section (d), that providers must comply with "examination, test and validation procedures to be carried out before, during and after the development of the high-risk AI system, and the frequency with which they have to be carried out". Without focusing here on the degree of risk of Information Retrieval or Classification systems, this Act clearly states that AI systems must tackle evolution. Time is a dimension that is often overlooked when conducting Information Retrieval (IR) experiments, especially when static data sets are utilized. The advantages of such datasets are that they are easily used to evaluate and test systems. Some data sets, like CORD19, contain documents collected at different points in time, showing differences in the set of documents from one collection time to another. Recent research <ref type="bibr" target="#b0">[1]</ref> has demonstrated that models trained on data pertaining to a particular time period struggle to keep their performance levels when applied on test data that is distant in time. On the other side, <ref type="bibr" target="#b1">[2]</ref> showed that neural systems, especially transformers-based ones, are not always very sensitive to corpus evolution.</p><p>With the aim of tackling this challenge of making models have persistent quality over time, the objective of the LongEval lab is twofold: (i) to explore the extent to which temporal differences over time, as reflected in the evolution of evaluation datasets, results in the deterioration of the performance of information retrieval and classification systems, and (ii) to propose improved methods that mitigate performance drop by making models more robust over time.</p><p>The LongEval lab <ref type="bibr" target="#b2">[3]</ref> took place as part of the Conference and Labs of the Evaluation Forum (CLEF) 2024, and consisted in two separate tasks: (i) Task 1, described in Section 2, focused on information retrieval, and (ii) Task 2, described in Section 3, focused on text classification for sentiment analysis. Both tasks provided labeled datasets enabling analysis and evaluation of models over data evolving in time (what we call "longitudinally evolving data"). In this paper, we add details to <ref type="bibr" target="#b3">[4]</ref>, by focusing on the datasets statistics, and on analysing in details the overall participant runs and results for each task.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Task 1 -Retrieval</head><p>The retrieval task of LongEval 2024 explores the effect of changes in datasets on retrieval of text documents. More specifically, we focus on a setup in which the datasets are evolving, as in the LongEval 2023 Retrieval Task data <ref type="bibr" target="#b2">[3]</ref>. This means, that one dataset can be acquired from another by adding, removing (and replacing) a limited number of documents and queries. The two main scenarios considered focus on one single system or on several ones, as detailed below:</p><p>A single system in an evolving setup We explore how one selected system behaves when evaluated on several collections, which evolve along the time. The context in which this task taked place is retrieval performances for Web search. When considering evolution of Web data along time, we are facing a case when the documents, the queries and also the relevance continuously evolves. We are then studying how Web search engines deal with this situation. The considered scenario is then similar to classical ad-hoc search, in the case of evolving data sets. The evaluation in this scenario consider both the Web search case in which the top documents are the most important elements considered, and should take into account the evolving nature of the data. Evaluation should ideally reflect the changes in the collection and especially signal substantial changes that could lead to performance drop. This would allow to re-train the search engine model then and only when it is really necessary, and enable much more efficient overall training.</p><p>As described earlier, there is no consensus about the stability of the performance of the neural networks IR systems along time, but it seems to be lower than in the case of statistical models. Moreover, the performance strongly depends on the data used for training the neural model. One objective of the task is to explore the behavior of the neural system in the evolving data scenario.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Comparison of multiple systems in an evolving setup</head><p>While in the first point, we explore a single system, comparison of this systems with multiple systems across evolving collections, should provide more information about systems stability and robustness.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1.">Description of the task</head><p>Compared to the LongEval 2023 Dataset <ref type="bibr" target="#b2">[3]</ref>, in 2024 we take larger lags between the training and the test sets. More precisely, the task is composed of:</p><p>• One training set, that contains Web documents, actual user's queries, and assessments, acquired at timestamp 𝑡; • Two test sets, acquired later than 𝑡 at time 𝑡 ′ and 𝑡", composed of Web documents and user's queries. The task datasets were created over sequential time periods, which allows doing observations at different time stamps 𝑡, and most importantly, comparing the performance across different time stamps 𝑡 and 𝑡 ′ . So, the IR task aims to assess the performance difference between 𝑡 ′ and 𝑡" when 𝑡 ′ occurs after 𝑡 ′ , according to teh fact that training set acquired at 𝑡, takes place few months before 𝑡 ′ .</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.">Dataset</head><p>As for LongEval 2023, in 2024 the data for this task were provided by the French search engine Qwant. They consist of the queries issued by the users of this search engine, cleaned Web documents, which were 1) selected to correspond to the queries, and 2) to add additional noise, and relevance judgments, which were created using a click model. The dataset is fully described in <ref type="bibr" target="#b4">[5]</ref>. We provided training data, which included 599 train queries, with corresponding 9,785 relevance assessments and 2,049,729 Web pages. All training data were collected during January 2023. The test set corpus is composed of two subsets: Lag6 acquired in June 2023 (i.e., 6 months later than the training set), and Lag8 acquired in August 2024 (i.e. acquired 8 months later than the training set). The test dataset contains 4,321,642 documents (June: 1,790,028; August: 2,531,614) and 1,925 test queries (June: 407; August: 1,518). The datasets are accessible through the lab's webpage<ref type="foot" target="#foot_1">2</ref> and from the TU Wien Research Data Repository <ref type="foot" target="#foot_2">3</ref> .</p><p>The data collected from the Qwant search engine is in French. In a way to help participants, the LongEval data set for the Retrieval task also contains automatic translations into English of both queries and documents. We mention however that the translations provided by LongEval are only applied to the first 500 characters of each sentence of the initial French documents downloaded.</p><p>The document and query overlap ratios between the collections is given by Table <ref type="table" target="#tab_0">1</ref> and Table <ref type="table" target="#tab_2">2</ref>. We see from these tables that there is a substantial overlap between the Train and the Test collection documents and (due to the larger size of the August query set) a substantial overlap between the Train / June queries and the August queries.  To evaluate the submissions we use one set of relevance judgments: the judgments acquired by the Qwant click model. For the evaluation, we use the NDCG measure (calculated for each dataset) at 10, as Ratio of the queries shared between the LongEval 2024 train and test collections, rows vs. columns, i.e. 0.99 means that 99% of queries in the row collection are also included in the column collection.  well as the drop between the Lag8 and Lag6 collection. This allows us to check to which extend the IR system face the evolution of the data. We also plan to use manual assessments, acquired through the interface described in section 2.8.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.3.">Submissions</head><p>14 teams submitted their systems to the Retrieval task. Each team was allowed to submit up to 10 systems. Together, this a overall of 73 runs submitted. Two teams submitted their runs on the wrong test data set, so we do not include their submission results in our further analysis.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.4.">Absolute Scores</head><p>For the Retrieval task of the LongEval lab, we computed two sets of scores for each of the lags in the test collection, namely NDCG and MAP. Table <ref type="table" target="#tab_4">3</ref> gives the overview of them for each run on the Lag6 and Lag8 datasets. For each run, the columns of the table indicate which language was used (English, French, or both), whether neural approaches were involved (values yes/no), and whether a single or a combination of several approaches was used (values yes/no). In addition, we show NDCG score histograms for these runs, in decreasing order, for each dataset, showing whether a run uses any neural approach (green for yes, yellow for no) in Figure <ref type="figure" target="#fig_0">1</ref>, and whether the run uses a combination of more than a single approach (orange for yes, cyan for no) in Figure <ref type="figure" target="#fig_2">2</ref>. This information was acquired from the participants through a questionnaire the participants had to fill for each submitted run. Figure <ref type="figure" target="#fig_3">3</ref> shows which language each made use of.</p><p>From Table <ref type="table" target="#tab_4">3</ref> we see that the systems which did best for the Lag6 data are also among the top for the Lag8, where the first ranked nine systems scores are comparable to each other. For instance, the best system on Lag6, according to the NDCG measure, (dam_run_4), is ranked the second best also on Lag8.</p><p>Similarity, the best system on Lag8, according to the NDCG measure, (mouse_run_8), is ranked the second best also on Lag6. This finding holds for the MAP measure as well.</p><p>Here, we describe the methods used in the top-3 runs, according to the NDCG evaluation measure, for both Lag6 and Lag8 datasets.   1. dam_run_4 from the DAM team: This system uses BM25 as a first stage retrieval model, enhanced with proximity search, query expansion via synonyms, and the MBNET model <ref type="bibr" target="#b5">[6]</ref>, which combines BERT and XLNET, for re-ranking the results.</p><p>2. mouse_run_8 from MOUSE team: This system also uses BM25 as a first stage retrieval model, enhanced with an LLM-based re-ranking model using the Cohere API <ref type="foot" target="#foot_3">4</ref> . It utilizes the Llama 3 model <ref type="bibr" target="#b6">[7]</ref> for query expansion.</p><p>3. mouse_run_10 from MOUSE team: Similar to mouse_run_8, this system uses BM25 as first stage retrieval model, but it is enhanced with a deep neural-based re-ranking model using PyGaggle. It also employs the Llama 3 model for query expansion.</p><p>For the Lag8 dataset, the top-3 systems are:</p><p>1. mouse_run_9 from MOUSE team: This system uses BM25 as a first stage retrieval model, enhanced with a deep neural-based re-ranking model using PyGaggle <ref type="foot" target="#foot_4">5</ref> . It uses the Mixtral model <ref type="bibr" target="#b7">[8]</ref> for query expansion.</p><p>2. mouse_run_8 from MOUSE team: Described above.</p><p>3. mouse_run_10 from MOUSE team: Described above.</p><p>Generally, most of the solutions chosen by the participants to the LongEval Retrieval task apply a multi-stage retrieval approach. Often, the first stage involves a lexical-based retrieval (e.g., BM25), and query expansion methods like PL2 or BO1. Query expansion is also done by employing Large Language Models, like Mistral or Llama 3. Reranking is done either using neural-based methods or sentence based transformers. Listwise rerankers and fusing have also been used in reranking of retrieved results. Notably, the temporal aspect of the LongEval test collection has been used by some participants to include past query relevance information into query reformulation either from clicklogs or from the documents deemed relevant in the previous Considering the Figures 1, 2 and 3, we see that the shape of the distribution of the NDCG values are similar for the Lag6 and Lag8 datasets. However, the systems have higher performances on Lag6 than on Lag8, with maximum 0.4 value for the NDCG on the Lag6 versus 0.3 for the Lag8.  <ref type="bibr" target="#b10">[11]</ref> yes yes French 0.392 0.293 0.244 0.171 mouse_run_9 <ref type="bibr" target="#b9">[10]</ref> yes yes French 0.392 0.298 0.245 0.175 iris_run_1 <ref type="bibr" target="#b10">[11]</ref> yes yes French 0.392 0.294 0.244 0.171 iris_run_2 <ref type="bibr" target="#b10">[11]</ref> yes yes French 0.392 0.293 0.242 0.170 iris_run_3 <ref type="bibr" target="#b10">[11]</ref> yes yes French 0.391 0.293 0.243 0.171 iris_run_5 <ref type="bibr" target="#b10">[11]</ref> yes French 0.390 0.294 0.240 0.171 mouse_run_7 <ref type="bibr" target="#b9">[10]</ref> yes no French 0.386 0.288 0.236 0.163 dam_run_3 <ref type="bibr" target="#b8">[9]</ref> no  <ref type="bibr" target="#b8">[9]</ref> yes no French 0.370 0.279 0.220 0.156 mouse_run_6 <ref type="bibr" target="#b9">[10]</ref> yes no French 0.367 0.286 0.215 0.162 cir_run_3 <ref type="bibr" target="#b11">[12]</ref> no no English 0.354 0.242 0.226 0.136 snu_run_1 <ref type="bibr" target="#b12">[13]</ref> yes yes English 0.334 0.251 0.197 0.142 ows_run_1 <ref type="bibr" target="#b12">[13]</ref> no no English 0.333 0.243 0.199 0.139 kalu_run_2 <ref type="bibr" target="#b13">[14]</ref> yes no French 0.330 0.254 0.192 0.143 kalu_run_3 <ref type="bibr" target="#b13">[14]</ref> yes no French 0.330 0.254 0.192 0.143 kalu_run_5 <ref type="bibr" target="#b13">[14]</ref> yes no Frencg 0.324 0.249 0.188 0.140 kalu_run_4 <ref type="bibr" target="#b13">[14]</ref> yes no French 0.323 0.250 0.186 0.140 cir_run_4 <ref type="bibr" target="#b11">[12]</ref> no no English 0.320 0.229 0.172 0.117 wonder_run_3 no no French,English 0.313 0.235 0.163 0.116 cir_run_2 <ref type="bibr" target="#b11">[12]</ref> yes no English 0.308 0.230 0.173 0.123 mouse_run_3 <ref type="bibr" target="#b9">[10]</ref> yes yes English 0.306 0.235 0.171 0.126 ows_run_2 <ref type="bibr" target="#b14">[15]</ref> no no English 0.306 0.229 0.197 0.140 dam_run_2 <ref type="bibr" target="#b8">[9]</ref> yes no English 0.304 0.231 0.169 0.121 mouse_run_4 <ref type="bibr" target="#b9">[10]</ref> yes yes English 0.304 0.232 0.167 0.124 mouse_run_5 <ref type="bibr" target="#b9">[10]</ref> yes yes English 0.304 0.232 0.166 0.124 wonder_run_4 no no French 0.299 0.223 0.155 0.107 kalu_run_1 <ref type="bibr" target="#b13">[14]</ref> no no French 0.298 0.219 0.158 0.107 galapagos_run_4 <ref type="bibr" target="#b15">[16]</ref> yes yes English 0.295 0.220 0.189 0.131 ows_run_3 <ref type="bibr" target="#b14">[15]</ref> yes yes English 0.294 0.224 0.188 0.135 dam_run_1 <ref type="bibr" target="#b8">[9]</ref> no no English 0.294 0.221 0.156 0.112 galapagos_run_5 <ref type="bibr" target="#b15">[16]</ref> yes yes English 0.293 0.221 0.187 0.132 mouse_run_2 <ref type="bibr" target="#b9">[10]</ref> yes no English 0.291 0.225 0.152 0.115 mouse_run_1 <ref type="bibr" target="#b9">[10]</ref> yes no English 0.291 0.225 0.153 0.114 ows_run_7 <ref type="bibr" target="#b14">[15]</ref> yes yes English 0.290 0.213 0.180 0.123 cir_run_5 <ref type="bibr" target="#b11">[12]</ref> no no English 0.285 0.212 0.148 0.104 ows_run_6 <ref type="bibr" target="#b14">[15]</ref> yes yes English 0.284 0.216 0.173 0.126 cir_run_1 <ref type="bibr" target="#b11">[12]</ref> no no English 0.282 0.211 0.145 0.103 snu_run_2 <ref type="bibr" target="#b12">[13]</ref> yes </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.5.">Changes in the Scores</head><p>The main part of the retrieval task is to study the changes in the performance scores between the collections. The collections were created using the same approach and procedure have a relatively high overlap in terms of both queries and documents (see Tables <ref type="table" target="#tab_2">1 and 2</ref>), we thus provide the Relative NDCG Drop (RND) values of systems between the collections Lag8 and Lag6. RnD(r) for a system 𝑟, is defined as as:</p><formula xml:id="formula_0">𝑅𝑁 𝐷(𝑟) = NDCG 𝐿𝑎𝑔6 (𝑟)−NDCG 𝐿𝑎𝑔8 (𝑟) NDCG 𝐿𝑎𝑔6 (𝑟)</formula><p>With such definition, small RND values man more robust systems against changes, and large RND values mean that the systems are not able to generalize well between lag6 and lag8. What we see in Table <ref type="table" target="#tab_7">4</ref> is that the systems which are more robust to the evolution of the test collections (low values on RND) are not the best ones: for instance, ows_run_4 is the more robust system but the third worse one in table 3. The best systems in term of NDCG values in lag6, 𝑑𝑎𝑚_𝑟𝑢𝑛 4 and 𝑚𝑜𝑢𝑠𝑒_𝑟𝑢𝑛_8, have an RND of 0.245, which means that they quite robust, but much less than the most robut ones. This shows that the very best systems do cope with some extend to the evolution of the corpus, but that their is room for improving best systems against robustness. We also see that the worse robust system against changes, cir_run_3, is a system that does not rely on neural IR models: such finding shows that neural models are also likely to be more robust against changes than non-neural ones. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.6.">Run Rankings</head><p>Another point of view studied is how the submitted runs compare to each other, either in terms of the absolute NDCG scores achieved on the collections, or in terms of NDCG changes between the collections. We also calculated the Pearson correlation between the runs (now shown here), with high correlation in terms of NDCG scores, 0.99, and similarly high, 0.98, with respect to ranking order. This corresponds to the relatively high overlaps of the documents and also the queries between Lag6 and Lag8 collections (Table <ref type="table" target="#tab_0">1</ref> and Table <ref type="table" target="#tab_2">2</ref>). This observation does not hold for the correlation between the ranking according to the NDCG score achieved and the ranking of the performance change, which is relatively low. The Pearson correlation is 0.07 for the Lag6 dataset and -0.05 on the Lag8 dataset. Last, we calculated a combination of both rankings (ranking in terms of absolute values and ranking in terms of change). For this, we first calculated a Borda count of the ranking in terms of absolute values and Borda count of the ranking in terms of relative change and then we simply summed these two Borda counts: this result is displayed in the last column in the Table <ref type="table" target="#tab_8">5</ref>. We see that in terms of this measure the top performing systems (on Lag6 and Lag8 datasets) are ranked higher, although they have lower rank in terms of the rank of the NDCG change. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.7.">Queries Overview</head><p>We further investigate performance on the provided queries. Due to the space reason, we only investigate a selected subset of queries from each collection. We used a pooling strategy to select these queries to be used for the manual assessment process (described in Section 2.8). We first selected the top five performing runs on the average NDCG performance on both collections. We then calculated the performance of these runs per queries for each collection (i.e. Lag6 and Lag8) and sorted the queries based on their NDCG performance for the five runs. Then, we divided the query set in each collection to four sets and randomly selected from each set: five and 10 queries from Lag 6 and Lag8, respectively. We selected in total 20 queries from Lag6 collection and 40 Lag8 collection. We selected more queries from Lag8 collection since, as shown in Table <ref type="table" target="#tab_2">2</ref>, the number of Lag8 collection is higher than Lag6 collection. Overview of the scores achieved for the selected queries in each collection is displayed in Figure <ref type="figure" target="#fig_5">4</ref>. The figure shows minimum performance (by any submitted run), 25%, quantile, 75% quantile and the maximum achieved NDCG score. Due to a relatively large number of runs, the range of the scores achieved is typically quite large and for some of the queries it even ranges between 0 and 0.8. It can be also noticed that the variation (corresponding to the size of the boxplot) of the query performance for the Lag8 collection is higher than Lag6 collection. Some of the worst performing queries are very general ("birdsong", "taxes", and "used car" for instance) and can thus be expected to be ambiguous. This is in contrast with the top performing queries (e.g. "camping concarneau", "Prune rabbit", and "point bordeaux vision") which refer to more specific information need. Some other top performing queries have high variation in the results, e.g. the query "origami bird" for which it is not specified if the user focuses about about "origami bird" or looks for tutorials to make them.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.8.">Manual relevance judgments acquisition</head><p>The evaluation results of LongEval IR task presented above rely on automatic assessments generated from click models <ref type="bibr" target="#b4">[5]</ref>. In addition to these click-based relevance assessments, we have set up an annotation tool to acquire further relevance assessments by humans. For that, we used the open source annotation tool, Doctag <ref type="bibr" target="#b16">[17]</ref>, on a sample of the queries selected in section 2.7 (60 queries in total). Doctag provides a customizable and portable platform specifically designed for Information Retrieval (IR) evaluation. To perform manual relevance judgments using Doctag, annotators utilize its web-based interface. They access the tool and interact with its annotation functionalities, including the assignment of labels to indicate document relevance to specific queries. Annotators view the documents and associate appropriate relevance labels (Fig. <ref type="figure" target="#fig_6">5</ref>). The documents to be annotated were selected through pooling the participants runs <ref type="bibr" target="#b17">[18]</ref>. For the annotation to remain tractable, we conducted a stratified   sampling and selected 60 queries for evaluation (Section 2.7). We set up dedicated online servers where Doctag is deployed, through their use we have acquired over 25K manual assessments. 2900 documents from the original dataset were then assessed. The average number of assessments per query is around 429. To perform the manual annotation and assess document relevance for the corresponding queries, we assigned subsets of the document dataset to a team of 25 annotators. We set up dedicated online servers where Doctag was deployed. Each annotator was assigned to a specific server to perform the annotation tasks. This distributed setup allowed for parallel processing, enabling annotators to work simultaneously and collaborate effectively within their assigned subsets.</p><p>We have recorded an aggregate of 25,759 judgments. These judgments span across four distinct categories: 'Relevant', 'Not Relevant', 'Partially Relevant', and 'I Don't Know'.</p><p>Preliminary analysis of the data indicates a more balanced approach among annotators in categorizing the query-document pairs. Figure <ref type="figure" target="#fig_7">6</ref> presents the judgment distribution for the top 30 queries in terms of document count. What we observe in Figure <ref type="figure" target="#fig_7">6</ref> is a more evenly distributed number of relevant (green) and non-relevant (red) documents for many queries. While some queries still show a high number of relevant documents (with peaks exceeding 300 relevant documents), the number of non-relevant documents is also significant, indicating no single dominant category. This balanced distribution of relevant and non-relevant documents is much more equitable than previous analyses, where nonrelevant judgments predominated.  The plots reveal that the distributions for relevant and not relevant judgments are similar, both with wide ranges and high densities around the median values.</p><p>Additionally, Figure <ref type="figure" target="#fig_8">7</ref> provides a detailed view of the distribution of judgment counts across all queries using violin plots. The violin plots reveal that the distributions for relevant and non-relevant judgments are quite similar, with both categories showing a wide range of counts and high densities around the median values. The partially relevant category, while also having a substantial number of judgments, shows a narrower distribution, indicating less variability. The "I don't know" category has a very narrow distribution, reflecting its infrequent use among annotators.</p><p>Further evaluation rounds utilizing the collected data are in progress. We will utilize the annotated documents and relevance annotations from the queries to construct an aggregated 𝑄𝑟𝑒𝑙 file. With this Qrel file, we will run the evaluation using trec_eval<ref type="foot" target="#foot_5">6</ref> on the participants' runs. Trec_eval will compare the system's retrieved results against the ground truth relevance judgments defined in the Qrel file. This evaluation process will provide valuable insights by comparing the results of the clic model with the manual annotations, thereby assessing the effectiveness and performance of the information retrieval system in relation to the specified queries.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.9.">Discussion and conclusion</head><p>This task was the second attempt to collectively investigate the impact of the evolution of the data on search system's performances. Having 14 participating teams submitting runs confirmed that this topic was of interest to the community.</p><p>The dataset released for this task consisted in a sequence of test collections corresponding to different times. The collections were composed of documents and queries coming from Qwant, and relevance judgment coming from a click model and manual assessment. While the manual assessment is ongoing at the time of the paper's publication, performances of participants' submitted runs were measured using the click logs.</p><p>Most of submitted runs rely on multi-stage retrieval approaches. In addition to the usage of Large Language Models in Query expansion. The effect of the translation of the documents and queries provided by the lab has a clear impact: the best results were obtained on the original French data.</p><p>Since each subset had substantial overlaps, the correlations between systems rankings was pretty high. As for the robustness of the systems towards dataset changes, we observed that the systems that are the more robust to the evolution of test collection were not the best performing ones.</p><p>Further evaluations will be carried out in the near future with the manual assessment of the pooled sets. A thorough analysis of the results will be necessary to study the impact of queries on the results (their nature, topic, difficulty, etc.). Further analysis work will be necessary to fully establish the robustness of the systems and the specific impact of dataset evolution on the performances.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Task 2 -Classification</head><p>Stance detection, an essential task in natural language processing (NLP), involves identifying an author's position or attitude towards a particular topic or statement. This task goes beyond simple sentiment analysis by requiring models to discern not just positive or negative sentiments but also the specific stance (supporting/believer, opposing/denier, or neutral) towards a given target <ref type="bibr" target="#b18">[19,</ref><ref type="bibr" target="#b19">20]</ref>.</p><p>Comprehending the evolution of social media stances over time poses a significant challenge, a topic that has gained recent interest in the AI and NLP communities but remains relatively unexplored. The performance of social media stance classifiers is intricately linked to temporal shifts in language and evolving societal attitudes toward the subject matter <ref type="bibr" target="#b20">[21]</ref>.</p><p>In LongEval 2024, social media stance detection, a multi-label English classification task, takes center stage, surpassing the complexity of the binary sentiment task in LongEval 2023 <ref type="bibr" target="#b21">[22]</ref>. Our primary goal is to assess the persistence of stance detection models in the dynamic landscape of social media posts.</p><p>The evolving nature of language and social opinions adds an additional layer of complexity to the challenges faced by text classifiers. Language undergoes continuous changes, reflecting shifts in societal norms and opinions and the emergence of novel concepts and words. For instance, consider the evolution of public opinion on climate change over the past two decades:</p><p>• Sentence from 2000: "Global warming is a theory that needs more proof; it's not urgent. "</p><p>• Sentence from 2010: "Evidence for climate change is mounting, and we need to start taking action. "</p><p>• Sentence from 2020: "Climate change is an undeniable crisis that requires immediate global action. "</p><p>The context over two decades in the above example shows that language and urgency surrounding climate change have evolved from skepticism to an accepted crisis. Models not updated with recent discussions and policy changes might fail to accurately capture the critical tone and terminology used in current dialogues about the environment. Similarly, the rapid emergence of new vocabulary, as witnessed with terms like COVID-19 <ref type="bibr" target="#b22">[23]</ref>, highlights the dynamic nature of language, presenting unique challenges for text classifiers.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.">Description of the task</head><p>To assess the extent of the performance drop of models over shorter and longer temporal gaps, we provided a comprehensive training dataset along with five testing sets. These testing sets include two practice sets and three development sets. The shared competition aimed to stimulate the development of classifiers that can effectively handle temporal variations and maintain performance persistence over different time distances. Participants were expected to submit solutions for two sub-tasks, showcasing their ability to address the challenges of temporal variations in performance. The shared task was in turn divided into two sub-tasks:</p><p>Sub-Task 1: Short-Term Persistence: In this sub-task, participants were tasked with developing models that demonstrated performance persistence over short periods. Specifically, the models needed to maintain their performance over a temporal gap between the within datasets and the short-term datasets. This involved comparing the performance from the within-practice data (January 2010 to December 2010) to the short-practice data (January 2014 to December 2014), a time gap of 4 years, and from the within-dev data (January 2011 to December 2011) to the short-dev data (January 2015 to December 2015), a time gap of 4 years Sub-Task 2: Long-Term Persistence: This sub-task required participants to develop models that maintained performance persistence over a longer period of time. The classifiers were expected to mitigate performance drops over a temporal gap between the within time datasets and the longterm datasets. This involved comparing the performance from the within-dev data (January 2011 to December 2011) to the long-dev data (January 2018 to September 2019), a time gap of approximately 7 to 8 years.</p><p>In addition to the main sub-tasks, participants were also asked to work on models that maintained performance within the same temporal year of the training set, with the practice-within data covering January 2010 to December 2010 and the within-dev data covering January 2011 to December 2011, with no gap between them and the training set (time gap 0).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.">Dataset</head><p>In this section, we present the process of constructing our final annotated corpus for the task. The large-scale Climate Change Twitter dataset was originally described in <ref type="bibr" target="#b23">[24]</ref>, Our primary focus will be on climate change stance, time of the post (created at), and the textual content of the tweets, which we will refer to as the CC-SD dataset. This CC-SD is large-scale, covering a span of 13 years and containing a diverse set of more than 15 million tweets from various years. Using the BERT model to annotated tweets, the CC-SD stance labels fall into three categories: those that express support for the belief in man-made climate change (believer), those that dispute it (denier), and those that remain neutral on the topic.</p><p>The total sum of the categorized tweets over the entire time span are as follows: 11,292,424 tweets as believers, 1,191,386 as deniers, and 3,305,601 as neutral, distributed across the timeline. The annotation is performed using transfer learning with BERT as distant supervision based on another sentiment climate change dataset <ref type="foot" target="#foot_6">7</ref> and, thus, can be easily manually annotated to improve its precision using human in the loop.</p><p>Data sampling. The dataset is first downsampled to ensure an equal number of instances for each stance (neutral, denier, believer) within a specified date range, using the minimum stance count across all selected months and years to avoid bias. This involves randomly sampling the same number of rows for each stance, year, and month combination, ensuring balanced representation. The downsampled data is then shuffled and split into training, development, and practice sets, including short-and long-term coverage, with any intersecting IDs between these sets being removed to maintain data integrity and prevent data leakage. Finally, a summary of the downsampled data is generated, detailing the number of rows, date and time of sampling, and statistics per year and month.</p><p>Test set annotation. We annotate our test data using Prolific<ref type="foot" target="#foot_7">8</ref> , which is a high quality data collection and annotation platform. The forms that contain data to annotate are created using Qualtrics <ref type="foot" target="#foot_8">9</ref> . We run the annotation in several batches, and provide the annotation guideline stating the task details and guidelines for the participants to follow. We add several filters, automatic and manual to select the optimal demographic and qualified annotators. Additionally, a manual annotation is also enforced which contains 5 tweets from the training set, which the organisers first annotate and then using the majority annotation is released as qualification task. The participant have to correctly answer 4 out of 5 questions to access the actual annotation task. We also provide fields in our form for every annotator to give their feedback and to point out if any tweet is inappropriate or contains explicit content in it. We collect responses from 5 annotators for each tweet, and select the majority annotation from the five annotation. In some cases, we find equal agreement among the annotators, and for those cases, we run an extra round of annotation to finalise the agreement. Finally after cleanup and majority annotation finding process, we manually check the data and divide into their respective splits.</p><p>The resulting distribution of data is shown in Table <ref type="table" target="#tab_9">6</ref>. table Dataset statistics summary of training, practice and testing sets. In the Practice phase, participants undertake Pre-Evaluation tasks with datasets from 2010 and 2014, sampled from CC-SD, allowing them to practice within a recent time frame and over a short duration. These datasets are manually verified. Additionally, human-annotated "within time" and "short time" practice sets are provided, also sampled from CC-SD, to refine model development before formal evaluation.</p><p>Subsequently, the Evaluation phase assesses models using datasets from 2011, 2015, and the longer period of 2018-2019, all sampled from CC-SD. These datasets undergo manual verification and encompass within-timeframe assessments, short-term predictions, and long-term predictions, offering a holistic evaluation of model performance across various temporal contexts. By incorporating datasets covering different years, the evaluation ensures thorough testing and understanding of models' temporal persistence and performance.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3.">Evaluation</head><p>Evaluation metrics for this edition of the task remain consistent with the previous version <ref type="bibr" target="#b2">[3,</ref><ref type="bibr" target="#b24">25]</ref>. All submissions were assessed using two key metrics: the macro-averaged F1-score on the corresponding sub-task's development set and the Relative Performance Drop (RPD), calculated by comparing performance on "within time" data against results from short-or long-term distant development sets. Submissions for each sub-task were ranked primarily based on the macro-averaged F1-score. Additionally, a unified score, the weighted-F1, was computed between the two sub-tasks, encouraging participants to contribute to both for accurate placement on a collective leaderboard and a deeper analysis of their system's performance in various settings.</p><p>Participants were expected to design an experimental architecture to enhance a text classifier's temporal performance. In such, the performance of the submissions was evaluated in two ways:</p><p>1. Macro-averaged F1-score: This metric measured the overall F1-score on the testing set for the sentiment classification sub-task. The F1-score combines precision and recall to provide a balanced measure of model performance. A higher F1-score indicated better performance in terms of both positive and negative sentiment classification.</p><formula xml:id="formula_1">𝐹 macro = 2 • precision • recall precision + recall<label>(1)</label></formula><p>2. Relative Performance Drop (RPD): This metric quantified the difference in performance between the "within-period" data and the short-or long-term distant testing sets. RPD was computed as the difference in performance scores between two sets. A negative RPD value indicated a drop in performance compared to the "within-period" data, while a positive value suggested an improvement.</p><formula xml:id="formula_2">𝑅𝑃 𝐷 = 𝑓 score𝑡 𝑗 − 𝑓 score𝑡 0 𝑓 score𝑡 0<label>(2)</label></formula><p>Where 𝑡 0 represents performance when the time gap is 0, and 𝑡 𝑗 represents performance when the time gap is short or long, as introduced in previous work <ref type="bibr" target="#b25">[26]</ref>.</p><p>The submissions were ranked primarily based on the macro-averaged F1-score, emphasizing the overall performance of the stance detection model on the testing sets. The higher the macro-averaged F1-score, the higher the ranking of the submission.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.4.">Models</head><p>In our study, we evaluated several baseline classifiers to assess their performance and temporal persistence when exposed to evolving data. The models we focused on include bert-base-uncased, roberta-base, and their respective variations with additional continual incremental pretraining from the climate change corpus.</p><p>To address the challenges posed by evolving data, we implemented continual incremental pretraining for both bert-base-uncased and roberta-base models. These variations, referred to as ++MLM 2019, were further pretrained on a climate change corpus that covers data from the initial training year up to 2019 using masked language modeling. This approach aimed to incorporate recent linguistic trends and contextual information, enhancing the models' ability to adapt to new and evolving data.</p><p>The dataset is segmented by years, starting from 2006 to various end years <ref type="bibr">(2011,</ref><ref type="bibr">2013,</ref><ref type="bibr">2015,</ref><ref type="bibr">2017,</ref><ref type="bibr">2019)</ref>. For each end year, data from all preceding years up to that point is aggregated and preprocessed. Preprocessing includes filling missing values with the most frequent value in each column, removing rows with missing values in the 'text' or 'stance' columns, and eliminating duplicate entries. Text data is normalized to lowercase, and entries with fewer than six words are excluded. Post-processing, the data is merged into a single dataset for each end year, resulting in five datasets representing different temporal spans. These datasets are subsequently balanced by downsampling to ensure uniform representation for incremental training.</p><p>Using a masked language modeling strategy, the textual data without its label is fed into the models incrementally in their chronological order, starting with the 2011 sample and ending with the 2019 sample. This approach ensures a balanced and clean dataset, facilitating robust analysis and model training. Each model was incrementally tested to evaluate its persistence over time, and the best performance was reported in the results section.</p><p>• bert-base-uncased (Bidirectional Encoder Representations from Transformers) <ref type="bibr" target="#b26">[27]</ref> is a foundational model in NLP that introduced the concept of bidirectional training of transformers for language modeling. The bert-base-uncased model is a version of BERT that ignores case sensitivity, which helps in learning case-independent features. It also consists of 12 transformer layers, 768 hidden units, and 12 attention heads. BERT uses a static masked language modeling objective during pretraining, which involves predicting masked words in a sentence based on their context.</p><p>• roberta-base (Robustly optimized BERT approach) <ref type="bibr" target="#b27">[28]</ref> is a variant of the BERT model designed to improve performance by optimizing the pretraining process. It uses dynamic masking, a larger batch size, and more data to enhance the training of transformer-based models. The roberta-base model consists of 12 transformer layers, 768 hidden units, and 12 attention heads. It is pretrained on a diverse range of data to capture rich contextual representations, making it effective for various NLP tasks.</p><p>• ++MLM 2019: A masked language modeling strategy used to adapt a language model to new data by incrementally pretraining with an unlabeled corpus up to 2019. This method leverages recent linguistic trends and contextual updates to improve model adaptation and performance over time.</p><p>This systematic approach allowed us to evaluate and enhance the models' temporal persistence and robustness baselines, ensuring they remain effective in the face of evolving language patterns.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.5.">Results</head><p>This section presents the results obtained during both the practice and evaluation phases of task 2.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.6.">Practice phase</head><p>In this subsection, we present the results of the practice phase of task 2. This practice dataset was provided to participants to allow them to practice and initiate their text classifiers. Since we did not get any submissions and to understand the initial performance of our practice sets, we compared several baseline classifiers. The models evaluated include roberta-base, bert-base-uncased, and their respective variations with additional continual incremental pretraining from the climate change corpus from the initial year of training up to 2019 using masked lanague modeling. The results are summarized in Table <ref type="table">7</ref>.</p><p>As it can be seen from Table <ref type="table">7</ref>, the results indicate that the ++MLM 2019 variations of both robertabase and bert-base-uncased demonstrate improved f-Within and f-Avg scores compared to their original counterparts. This suggests that additional continual pretraining based on recent data, incrementally over time, contributes to better performance persistence. Notably, bert-base-uncased ++MLM 2019 achieved the lowest RPD, highlighting its resilience to temporal changes.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 7</head><p>Performance of baseline models on practice data. The columns represent: f-Within -performance within the same time period, f-Short -performance over short temporal gaps, f-Avg -average performance across all temporal gaps, and RPD -relative performance drop when applied to temporally distant data. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Model</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.7.">Evaluation phase</head><p>In this subsection, we present the results of the evaluation phase of task 2. Using the development dataset provided to participants, we evaluated the final performance of the text classifier models. To understand the performance of our development sets, we compared several baseline classifiers due to the lack of submissions. The models evaluated include roberta-base, bert-base-uncased, and their respective variations with additional continual incremental pretraining from the climate change corpus up to 2019 using masked language modeling. The results are summarized in Table <ref type="table">8</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 8</head><p>Performance of baseline models on development sets. The columns represent: f-Within -performance within the same time period, f-Short -performance over short temporal gaps, f-Long -performance over long temporal gaps, f-Avg -average performance across all temporal gaps, RPD-Short -relative performance drop over short temporal gaps, RPD-Long -relative performance drop over long temporal gaps, and RPD-Avg -average relative performance drop. As shown in Table <ref type="table">8</ref>, the ++MLM 2019 variations of both roberta-base and bert-base-uncased models exhibit notable improvements in the f-Short and f-Long scores, as well as reduced RPD values compared to their standard counterparts. The ++MLM 2019 variation of roberta-base achieved an f-Avg score of (0.590), an improvement over the original model's score of (0.571). It also showed a significantly lower RPD-Short of (-4.74%) and RPD-Long of (-11.46%), indicating better resilience to temporal changes over both short and long gaps. Similarly, the ++MLM 2019 variation of bert-base-uncased achieved an f-Avg score of (0.570), slightly lower than the original model's 0.573. However, it exhibited a lower RPD-Long of (-10.01%) and RPD-Avg of (-14.94%), demonstrating improved performance persistence over time.</p><p>These results reinforce the value of continual incremental pretraining with recent data to maintain and improve model performance in dynamic environments. The ++MLM 2019 variations consistently showed enhanced performance metrics and reduced performance degradation over time, validating the effectiveness of this approach in enhancing temporal persistence.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.8.">Discussion and conclusion</head><p>This section discusses the results of our study on temporally adaptive classification methods, highlighting the significance of incorporating temporal information into text classification models to mitigate performance drops over time and the use of an outdated language model. These results reveal that classifiers trained on older data exhibit significant performance drops when applied to newer data. This is evident from the relative performance drops (RPD) reported, where the ++MLM 2019 variations showed a marked improvement in mitigating this drop.</p><p>Previous work by Alkhalifa et al. <ref type="bibr" target="#b25">[26]</ref> introduced the Incremental Temporal Alignment (ITA) method as a superior approach for enhancing temporal persistence of static word embedding. This method aligns closely with the continual incremental pretraining approach evaluated in our results, where ++MLM 2019 variations of both roberta-base and bert-base-uncased demonstrated improved f-Within, f-Avg scores, and lower RPD values. The ITA method's emphasis on leveraging incremental updates to word embeddings aligns with the improvements seen in the ++MLM 2019 models, showcasing their resilience to evolving data and enhancing their persistence as text classifiers as context updated overtime.</p><p>The results reinforce several best practices for designing temporally robust and persistent text classifiers. Methods relying on incremental updates generally outperform static embeddings, as corroborated by the superior performance of the ++MLM 2019 models. Additionally, it is crucial to select robust baseline models and incrementally update them to accommodate evolving language patterns over time.</p><p>The practical implications of our findings are significant for real-world NLP applications. In dynamic environments such as stance posts on social media, language evolves rapidly, making temporal adaptation through an incremental pretraining approach substantially enhance the longevity and persistence of text classifiers. These results provide empirical evidence supporting the implementation of temporally adaptive classification methods in real-world scenarios.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Overview of the systems using a neural approach (green) vs. other (yellow).</figDesc><graphic coords="4,72.00,171.68,225.64,128.94" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: Overview of the systems which use a single approach (orange) and which use a combination of multiple approaches (cyan)</figDesc><graphic coords="5,72.00,257.87,225.64,128.94" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Figure 3 :</head><label>3</label><figDesc>Figure 3: Overview of the systems which use French (blue), which use English translations (red), and which use both (purple).</figDesc><graphic coords="5,297.64,257.87,225.64,128.94" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_5"><head>Figure 4 :</head><label>4</label><figDesc>Figure 4: Selected queries performance from Lag6 and Lag8 datasets.</figDesc><graphic coords="12,72.00,273.39,451.26,195.19" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_6"><head>Figure 5 :</head><label>5</label><figDesc>Figure 5: Screenshot from Doctag main page. Labels annotation is done associating to each document one label that expresses the relevance of that document for that topic.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_7"><head>Figure 6 :</head><label>6</label><figDesc>Figure 6: The distribution of judgment votes for the top 30 queries based on document count. Resulting counts of 'Relevant' (green), 'Not Relevant' (red), and 'Partially Relevant' (orange) votes are shown.</figDesc><graphic coords="13,94.57,69.31,406.15,304.39" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_8"><head>Figure 7 :</head><label>7</label><figDesc>Figure 7: Violin plots showing the distribution of judgment counts across different categories for all queries.The plots reveal that the distributions for relevant and not relevant judgments are similar, both with wide ranges and high densities around the median values.</figDesc><graphic coords="13,94.57,422.53,406.15,295.10" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1</head><label>1</label><figDesc>Ratio of documents shared between the LongEval 2024 train and test collections, row vs. column, i.e. 0.93 means that 93% of documents in the row collection are also included in the column collection.</figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 2</head><label>2</label><figDesc></figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Train 2024 June (Lag6) August (Lag8)</head><label></label><figDesc></figDesc><table><row><cell>Train 2024</cell><cell>1.00</cell><cell>0.22</cell><cell>0.42</cell></row><row><cell>June (Lag6)</cell><cell>0.32</cell><cell>1.00</cell><cell>0.56</cell></row><row><cell>August (Lag8)</cell><cell>0.17</cell><cell>0.15</cell><cell>1.00</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_4"><head>Table 3 :</head><label>3</label><figDesc>NDCG and MAP scores for Lag6, Lag8. Results are sorted according to the NDCG scores on the Lag6.</figDesc><table><row><cell>NDCG</cell><cell>MAP</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_6"><head>318 0.238 0.183 0.129</head><label></label><figDesc></figDesc><table><row><cell></cell><cell></cell><cell>yes</cell><cell>English 0.282 0.213 0.177 0.127</cell></row><row><cell>lfzzo_run_4</cell><cell>no</cell><cell>no</cell><cell>English 0.280 0.209 0.142 0.102</cell></row><row><cell>lfzzo_run_2</cell><cell>no</cell><cell>no</cell><cell>English 0.280 0.207 0.142 0.099</cell></row><row><cell>wonder_run_2</cell><cell>no</cell><cell>no</cell><cell>English 0.279 0.207 0.137 0.099</cell></row><row><cell>lfzzo_run_3</cell><cell>no</cell><cell>no</cell><cell>English 0.277 0.209 0.139 0.102</cell></row><row><cell>lfzzo_run_1</cell><cell>no</cell><cell>no</cell><cell>English 0.276 0.207 0.140 0.100</cell></row><row><cell>lfzzo_run_5</cell><cell>no</cell><cell>no</cell><cell>English 0.274 0.207 0.137 0.101</cell></row><row><cell>seekx_run_1</cell><cell>no</cell><cell>no</cell><cell>French 0.274 0.201 0.145 0.095</cell></row><row><cell>seekx_run_2</cell><cell>no</cell><cell>no</cell><cell>French 0.274 0.202 0.144 0.096</cell></row><row><cell>seekx_run_4</cell><cell>no</cell><cell>no</cell><cell>English 0.273 0.202 0.139 0.098</cell></row><row><cell>wonder_run_5</cell><cell>no</cell><cell>no</cell><cell>English 0.273 0.203 0.137 0.098</cell></row><row><cell>wonder_run_1</cell><cell>no</cell><cell>no</cell><cell>English 0.272 0.203 0.136 0.098</cell></row><row><cell>seekx_run_5</cell><cell>no</cell><cell>no</cell><cell>English 0.264 0.193 0.133 0.091</cell></row><row><cell>galapagos_run_2 [16]</cell><cell>yes</cell><cell>yes</cell><cell>English 0.261 0.198 0.162 0.115</cell></row><row><cell>galapagos_run_1 [16]</cell><cell>yes</cell><cell>yes</cell><cell>English 0.258 0.196 0.157 0.111</cell></row><row><cell>galapagos_run_3 [16]</cell><cell>yes</cell><cell>yes</cell><cell>English 0.253 0.192 0.151 0.107</cell></row><row><cell>ows_run_4 [15]</cell><cell>yes</cell><cell>yes</cell><cell>English 0.246 0.204 0.128 0.114</cell></row><row><cell>ows_run_5 [15]</cell><cell>no</cell><cell>yes</cell><cell>English 0.240 0.177 0.124 0.085</cell></row><row><cell>seekx_run_3</cell><cell>no</cell><cell>no</cell><cell>French 0.236 0.174 0.120 0.079</cell></row><row><cell>AVERAGE</cell><cell></cell><cell></cell><cell>0.</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_7"><head>Table 4 :</head><label>4</label><figDesc>Changes in the NDCG scores. Lines are ordered by descending RND values.</figDesc><table><row><cell></cell><cell>NDCG</cell><cell>RND</cell></row><row><cell>System</cell><cell>Lag6 Lag8</cell></row><row><cell>ows_run_4</cell><cell cols="2">0.246 0.204 0.169</cell></row><row><cell>mouse_run_6</cell><cell cols="2">0.367 0.286 0.220</cell></row><row><cell>kalu_run_4</cell><cell cols="2">0.323 0.250 0.224</cell></row><row><cell>mouse_run_1</cell><cell cols="2">0.291 0.225 0.226</cell></row><row><cell>mouse_run_2</cell><cell cols="2">0.291 0.225 0.229</cell></row><row><cell>kalu_run_2</cell><cell cols="2">0.330 0.254 0.230</cell></row><row><cell>kalu_run_5</cell><cell cols="2">0.324 0.249 0.230</cell></row><row><cell>mouse_run_3</cell><cell cols="2">0.306 0.235 0.231</cell></row><row><cell>kalu_run_3</cell><cell cols="2">0.330 0.254 0.232</cell></row><row><cell>mouse_run_5</cell><cell cols="2">0.304 0.232 0.235</cell></row><row><cell>mouse_run_4</cell><cell cols="2">0.304 0.232 0.235</cell></row><row><cell>ows_run_6</cell><cell cols="2">0.284 0.216 0.238</cell></row><row><cell cols="3">galapagos_run_1 0.258 0.196 0.239</cell></row><row><cell>ows_run_3</cell><cell cols="2">0.294 0.224 0.239</cell></row><row><cell>mouse_run_9</cell><cell cols="2">0.392 0.298 0.240</cell></row><row><cell cols="3">galapagos_run_2 0.261 0.198 0.241</cell></row><row><cell>dam_run_2</cell><cell cols="2">0.304 0.231 0.241</cell></row><row><cell>mouse_run_10</cell><cell cols="2">0.393 0.298 0.243</cell></row><row><cell cols="3">galapagos_run_3 0.253 0.192 0.243</cell></row><row><cell>lfzzo_run_3</cell><cell cols="2">0.277 0.209 0.243</cell></row><row><cell>snu_run_2</cell><cell cols="2">0.282 0.213 0.245</cell></row><row><cell>mouse_run_8</cell><cell cols="2">0.395 0.298 0.245</cell></row><row><cell>dam_run_5</cell><cell cols="2">0.370 0.279 0.245</cell></row><row><cell>lfzzo_run_5</cell><cell cols="2">0.274 0.207 0.245</cell></row><row><cell>wonder_run_3</cell><cell cols="2">0.313 0.235 0.247</cell></row><row><cell>iris_run_5</cell><cell cols="2">0.390 0.294 0.248</cell></row><row><cell cols="3">galapagos_run_5 0.293 0.221 0.248</cell></row><row><cell>dam_run_1</cell><cell cols="2">0.294 0.221 0.249</cell></row><row><cell>snu_run_1</cell><cell cols="2">0.334 0.251 0.250</cell></row><row><cell>iris_run_3</cell><cell cols="2">0.391 0.293 0.251</cell></row><row><cell>lfzzo_run_1</cell><cell cols="2">0.276 0.207 0.251</cell></row><row><cell>ows_run_2</cell><cell cols="2">0.306 0.229 0.251</cell></row><row><cell>iris_run_2</cell><cell cols="2">0.392 0.293 0.251</cell></row><row><cell>iris_run_1</cell><cell cols="2">0.392 0.294 0.251</cell></row><row><cell>lfzzo_run_4</cell><cell cols="2">0.280 0.209 0.252</cell></row><row><cell>iris_run_4</cell><cell cols="2">0.392 0.293 0.252</cell></row><row><cell>cir_run_2</cell><cell cols="2">0.308 0.230 0.252</cell></row><row><cell>cir_run_1</cell><cell cols="2">0.282 0.211 0.252</cell></row><row><cell>wonder_run_1</cell><cell cols="2">0.272 0.203 0.253</cell></row><row><cell>wonder_run_4</cell><cell cols="2">0.299 0.223 0.253</cell></row><row><cell>mouse_run_7</cell><cell cols="2">0.386 0.288 0.255</cell></row><row><cell cols="3">galapagos_run_4 0.295 0.220 0.256</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_8"><head>Table 5 :</head><label>5</label><figDesc>Ranking of the submitted systems by NDCG scores (columns 2-3), changes in NDCG scores between Lag6 and Lag8 dataset (column 4). Column 4 shows the sum of the Borda count applied to ranking on Lag6 and Lag8 datasets and Borda count of ranking change between Lag8 and Lag6 dataset. The darker color means better performance.</figDesc><table><row><cell>System</cell><cell>NDCG Lag6</cell><cell>NDCG Lag8</cell><cell>RND</cell><cell>Borda</cell></row><row><cell>dam_run_4</cell><cell>1</cell><cell>4</cell><cell>45</cell><cell>151</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_9"><head>Table 6</head><label>6</label><figDesc>Dataset statistics summary of training, practice and testing sets.</figDesc><table><row><cell>Dataset</cell><cell>Time Period</cell><cell>Size</cell></row><row><cell>train</cell><cell cols="2">January 2009 to December 2011 35739</cell></row><row><cell cols="2">within-practice January 2010 to December 2010</cell><cell>450</cell></row><row><cell>short-practice</cell><cell>January 2014 to December 2014</cell><cell>450</cell></row><row><cell>dev-within</cell><cell>January 2011 to December 2011</cell><cell>1074</cell></row><row><cell>dev-short</cell><cell>January 2015 to December 2015</cell><cell>1074</cell></row><row><cell>dev-long</cell><cell cols="2">January 2018 to September 2019 1074</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">https://www.europarl.europa.eu/doceo/document/TA-9-2024-0138_EN.html</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_1">https://clef-longeval.github.io/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_2">https://doi.org/10.48436/xr350-79683</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_3">https://docs.cohere.com/docs/rerank-2</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_4">https://github.com/castorini/pygaggle</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="6" xml:id="foot_5">https://trec.nist.gov/trec_eval/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="7" xml:id="foot_6">https://www.kaggle.com/datasets/edqian/twitter-climate-change-sentiment-dataset</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="8" xml:id="foot_7">https://www.prolific.com/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="9" xml:id="foot_8">https://www.qualtrics.com/</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgments</head><p>This work is supported by the ANR Kodicare bi-lateral project, grant ANR-19-CE23-0029 of the French Agence Nationale de la Recherche, and by the Austrian Science Fund (FWF, grant I4471-N). This work is also supported by a UKRI/EPSRC Turing AI Fellowship to Maria Liakata (grant no. EP/V030302/1). This work has been using services provided by the LINDAT/CLARIAH-CZ Research Infrastructure (https://lindat.cz), supported by the Ministry of Education, Youth and Sports of the Czech Republic (Project No. LM2023062) and has been also supported by the Ministry of Education, Youth and Sports of the Czech Republic, Project No. LM2023062 LINDAT/CLARIAH-CZ.</p></div>
			</div>


			<div type="availability">
<div xmlns="http://www.tei-c.org/ns/1.0"><p>dam_run_1 seupd2324-dam_EN-Stop-SnowBall-Poss-Prox(50) dam_run_2 seupd2324-dam_EN-Stop-SnowBall-Poss-Prox(50)-Reranking(200) dam_run_3 seupd2324-dam_FR-Stop-FrenchLight-Elision-ICU-Prox(50) dam_run_4 seupd2324-dam_FR-Stop-FrenchLight-Elision-ICU-Prox(50)-Reranking(150) dam_run_5 seupd2324-dam_FR-Stop-FrenchLight-Elision-ICU-Shingles-Prox(50)-Reranking(150) iris_run_1</p><p>seupd2324-iris_FR_GFF@12_w0.162_MMARCO@1000_ADD_w5 iris_run_2 seupd2324-iris_FR_GFF@12_w0.162_MMARCO@1000_MAXMIN_ADD_w5 iris_run_3 seupd2324-iris_FR_MMARCO@1000_ADD_w5 iris_run_4 seupd2324-iris_FR_url_w1.4_GFF@12_w0.162_MMARCO@1000_ADD_w5 iris_run_5 seupd2324-iris-FR_Q2K@1_w0.16_MMARCO@1000_MAXMIN_ADD_w5 lfzzo_run_1 seupd2324-lfzzo-englishSystem1 lfzzo_run_2 seupd2324-lfzzo-englishSystem2 lfzzo_run_3 seupd2324-lfzzo-englishSystem3 lfzzo_run_4 seupd2324-lfzzo-englishSystem4 lfzzo_run_5 seupd2324-lfzzo-englishSystem5 lfzzo_run_6 seupd2324-lfzzo-frenchSystem1 lfzzo_run_7 seupd2324-lfzzo-frenchSystem2 lfzzo_run_8 seupd2324-lfzzo-frenchSystem3 lfzzo_run_9 seupd2324-lfzzo-frenchSystem4 lfzzo_run_10 seupd2324-lfzzo-frenchSystem5 mouse_run_1 seupd2324-mouse_English_Porter_Standard_NoStop_Mixtral-8x7b_NoRerank mouse_run_2 seupd2324-mouse_English_Porter_Standard_stopwords-en_LLama3-70b_NoRerank mouse_run_3 seupd2324-mouse_English_Porter_Standard_top125_LLama3-70b_Cohere-100-w06 mouse_run_4 seupd2324-mouse_English_Porter_Standard_top125_LLama3-70b_Pygaggle-Luyu-20-w06 mouse_run_5 seupd2324-mouse_English_Porter_Standard_top125_Mixtral-8x7b_Pygaggle-Luyu-20-w06 mouse_run_6 seupd2324-mouse_French_FrenchLight_Standard_NoStop_Mixtral-8x7b_NoRerank mouse_run_7 seupd2324-mouse_French_FrenchLight_Standard_stopwords-fr_LLama3-70b_NoRerank mouse_run_8 seupd2324-mouse_French_FrenchLight_Standard_top125_LLama3-70b_Cohere-100-w06 mouse_run_9 seupd2324-mouse_French_FrenchLight_Standard_top125_LLama3-70b_Pygaggle-Luyu-20-w06 mouse_run_10 seupd2324-mouse_French_FrenchLight_Standard_top125_Mixtral-8x7b_Pygaggle-Luyu-20-w06 seekx_run_1 seupd2324-seekx_LetLightFR seekx_run_2 seupd2324-seekx_LetLightStopFR seekx_run_3 seupd2324-seekx_LetLightStopSynFR seekx_run_4 seupd2324-seekx_StanMinEN seekx_run_5 seupd2324-seekx_StanMinSynEN snu_run_1 SNU_LDI_listt5 snu_run_2 SNU_LDI_monot5 wonder_run_1 WONDER_BASELINE wonder_run_2 WONDER_ENGLISH wonder_run_3 WONDER_ENGLISH_FRENCH wonder_run_4 WONDER_FRENCH wonder_run_5 WONDER_TWOPHASE xplore_run_1 XPLORE_French-BM25-FrenchLight-Stop xplore_run_2 XPLORE_French-BM25-FrenchLight-Stop-SynonymMapper xplore_run_3 XPLORE_French-BM25Default-FrenchLight-Stop xplore_run_4 XPLORE_French-LMDirichlet-FrenchLight-Stop</p></div>
			</div>

			<div type="annex">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>A. Runs submitted to the IR Task Table 9</head><p>The original name of the submitted runs for the IR task are shown in the second column while the Runs Ids used assigned to the systems and used in the paper are shown in the first column.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Run Id</head><p>Submitted System </p></div>			</div>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Synthetic target domain supervision for open retrieval qa</title>
		<author>
			<persName><forename type="first">R</forename><surname>Gangi Reddy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Iyer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">A</forename><surname>Sultan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Sil</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Castelli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Florian</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Roukos</surname></persName>
		</author>
		<idno type="DOI">10.1145/3404835.3463085</idno>
		<idno>doi:10.1145/3404835.3463085</idno>
		<ptr target="https://doi.org/10.1145/3404835.3463085" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR &apos;21</title>
				<meeting>the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR &apos;21<address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>Association for Computing Machinery</publisher>
			<date type="published" when="2021">2021</date>
			<biblScope unit="page" from="1793" to="1797" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<monogr>
		<author>
			<persName><forename type="first">J</forename><surname>Lovón-Melgarejo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Soulier</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Pinel-Sauvagnat</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Tamine</surname></persName>
		</author>
		<idno type="DOI">10.1007/978-3-030-72113-8_25</idno>
		<idno>doi:</idno>
		<ptr target="10.1007/978-3-030-72113-8_25" />
		<title level="m">Studying catastrophic forgetting in neural ranking models</title>
				<meeting><address><addrLine>Berlin, Heidelberg</addrLine></address></meeting>
		<imprint>
			<publisher>Springer-Verlag</publisher>
			<date type="published" when="2021">2021</date>
			<biblScope unit="page" from="375" to="390" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Overview of the clef-2023 longeval lab on longitudinal evaluation of model performance</title>
		<author>
			<persName><forename type="first">R</forename><surname>Alkhalifa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Bilal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Borkakoty</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Camacho-Collados</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Deveaud</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>El-Ebshihy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Espinosa-Anke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Gonzalez-Saez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Galuščáková</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Goeuriot</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Kochkina</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Liakata</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Loureiro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><forename type="middle">T</forename><surname>Madabushi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Mulhem</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Piroi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Popel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Servan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Zubiaga</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Experimental IR Meets Multilinguality, Multimodality, and Interaction. Proceedings of the Fourteenth International Conference of the CLEF Association (CLEF 2023)</title>
		<title level="s">Lecture Notes in Computer Science (LNCS</title>
		<meeting><address><addrLine>Thessaliniki, Greece</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Overview of the CLEF 2024 LongEval Lab on Longitudinal Evaluation of Model Performance</title>
		<author>
			<persName><forename type="first">R</forename><surname>Alkhalifa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Borkakoty</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Deveaud</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>El-Ebshihy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Espinosa-Anke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Fink</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Galuščáková</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Gonzalez-Saez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Goeuriot</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Iommi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Liakata</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><forename type="middle">T</forename><surname>Madabushi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Medina-Alias</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Mulhem</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Piroi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Popel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Zubiaga</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Experimental IR Meets Multilinguality, Multimodality, and Interaction. Proceedings of the Fifteenth International Conference of the CLEF Association (CLEF 2024)</title>
		<title level="s">Lecture Notes in Computer Science (LNCS</title>
		<editor>
			<persName><forename type="first">L</forename><surname>Goeuriot</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Mulhem</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><surname>Quénot</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">D</forename><surname>Schwab</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">L</forename><surname>Soulier</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><forename type="middle">M D</forename><surname>Nunzio</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Galuščáková</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><forename type="middle">G S</forename><surname>De Herrera</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><surname>Faggioli</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</editor>
		<meeting><address><addrLine>Heidelberg, Germany</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<monogr>
		<author>
			<persName><forename type="first">P</forename><surname>Galuščáková</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Deveaud</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Gonzalez-Saez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Mulhem</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Goeuriot</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Piroi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Popel</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2303.03229</idno>
		<title level="m">Longevalretrieval: French-english dynamic test collection for continuous web search evaluation</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Mpnet: Masked and permuted pre-training for language understanding</title>
		<author>
			<persName><forename type="first">K</forename><surname>Song</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Tan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Qin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Lu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T.-Y</forename><surname>Liu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Advances in neural information processing systems</title>
		<imprint>
			<biblScope unit="volume">33</biblScope>
			<biblScope unit="page" from="16857" to="16867" />
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<monogr>
		<author>
			<persName><forename type="first">H</forename><surname>Touvron</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Lavril</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Izacard</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Martinet</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M.-A</forename><surname>Lachaux</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Lacroix</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Rozière</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Goyal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Hambro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Azhar</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2302.13971</idno>
		<title level="m">Llama: Open and efficient foundation language models</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b7">
	<monogr>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">Q</forename><surname>Jiang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Sablayrolles</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Roux</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Mensch</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Savary</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Bamford</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">S</forename><surname>Chaplot</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Casas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">B</forename><surname>Hanna</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Bressand</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2401.04088</idno>
		<title level="m">Mixtral of experts</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b8">
	<monogr>
		<title level="m" type="main">Seupd@clef: Team dam on reranking using sentence embedders</title>
		<author>
			<persName><forename type="first">A</forename><surname>Basaglia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Stocco</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Popović</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<monogr>
		<title level="m" type="main">Seupd@clef: Team mouse on enhancing search engines effectiveness with large language models</title>
		<author>
			<persName><forename type="first">L</forename><surname>Cazzador</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">L D</forename><surname>Faveri</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Franceschini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Pamio</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Piron</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<monogr>
		<title level="m" type="main">Seupd@clef: Team iris on temporal evolution of query expansion and rank fusion techniques applied to cross-encoder re-rankers</title>
		<author>
			<persName><forename type="first">F</forename><surname>Galli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Rigobello</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Schibuola</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Zuech</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<monogr>
		<title level="m" type="main">Leveraging prior relevance signals in web search</title>
		<author>
			<persName><forename type="first">J</forename><surname>Keller</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Breuer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Schaer</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<monogr>
		<title level="m" type="main">Analyzing the effectiveness of listwise reranking with positional invariance on temporal generalizability</title>
		<author>
			<persName><forename type="first">S</forename><surname>Yoon</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Kim</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Hwang</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<monogr>
		<title level="m" type="main">Seupd@clef: Team kalu on improving search engine performance with query expansion and re-ranking approach</title>
		<author>
			<persName><forename type="first">A</forename><surname>Kimia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Akan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Arwa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<author>
			<persName><forename type="first">D</forename><surname>Alexander</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Fröbe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Hendriksen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Schlatt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Hagen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Hiemstra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Potthast</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">P</forename><surname>Vries</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Team openwebsearch at clef 2024</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<monogr>
		<title level="m" type="main">Team galápagos tortoise at longeval 2024: Neural re-ranking and rank fusion for temporal stability</title>
		<author>
			<persName><forename type="first">M</forename><surname>Gründel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Weber</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Franke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">H</forename><surname>Reimer</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Doctag: A customizable annotation tool for ground truth creation</title>
		<author>
			<persName><forename type="first">F</forename><surname>Giachelle</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Irrera</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Silvello</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Advances in Information Retrieval: 44th European Conference on IR Research, ECIR 2022</title>
		<title level="s">Lecture Notes in Computer Science</title>
		<meeting><address><addrLine>Stavanger, Norway</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2022">April 10-14, 2022. 2022</date>
			<biblScope unit="volume">13186</biblScope>
			<biblScope unit="page" from="288" to="293" />
		</imprint>
	</monogr>
	<note>Proceedings, Part II</note>
</biblStruct>

<biblStruct xml:id="b17">
	<monogr>
		<author>
			<persName><forename type="first">D</forename><surname>Harman</surname></persName>
		</author>
		<idno type="DOI">10.1007/978-3-642-36415-0_7</idno>
		<ptr target="https://doi.org/10.1007/978-3-642-36415-0_7.doi:10.1007/978-3-642-36415-0_7" />
		<title level="m">TREC-Style Evaluations</title>
				<meeting><address><addrLine>Berlin Heidelberg; Berlin, Heidelberg</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2013">2013</date>
			<biblScope unit="page" from="97" to="115" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">Stance detection: A survey</title>
		<author>
			<persName><forename type="first">D</forename><surname>Küçük</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Can</surname></persName>
		</author>
		<idno type="DOI">10.1145/3369026</idno>
		<ptr target="https://doi.org/10.1145/3369026.doi:10.1145/3369026" />
	</analytic>
	<monogr>
		<title level="j">ACM Comput. Surv</title>
		<imprint>
			<biblScope unit="volume">53</biblScope>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">Stance and sentiment in Tweets</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">M</forename><surname>Mohammad</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Sobhani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Kiritchenko</surname></persName>
		</author>
		<idno type="DOI">10.1145/3003433</idno>
		<ptr target="http://alt.qcri.org/semeval2016/task6/.doi:10.1145/3003433.arXiv:1605.01655" />
	</analytic>
	<monogr>
		<title level="j">ACM Transactions on Internet Technology</title>
		<imprint>
			<biblScope unit="volume">17</biblScope>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">Capturing stance dynamics in social media: open challenges and research directions</title>
		<author>
			<persName><forename type="first">R</forename><surname>Alkhalifa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Zubiaga</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">International Journal of Digital Humanities</title>
		<imprint>
			<biblScope unit="page" from="1" to="21" />
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">Longeval: Longitudinal evaluation of model performance at clef</title>
		<author>
			<persName><forename type="first">R</forename><surname>Alkhalifa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Bilal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Borkakoty</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Camacho-Collados</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Deveaud</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>El-Ebshihy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Espinosa-Anke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Gonzalez-Saez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Galuščáková</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Goeuriot</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Kochkina</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Liakata</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Loureiro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Tayyar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Madabushi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Mulhem</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Piroi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Popel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Servan</surname></persName>
		</author>
		<author>
			<persName><surname>Zubiaga</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Advances in Information Retrieval</title>
				<editor>
			<persName><forename type="first">J</forename><surname>Kamps</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">L</forename><surname>Goeuriot</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">F</forename><surname>Crestani</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><surname>Maistro</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">H</forename><surname>Joho</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">B</forename><surname>Davis</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">C</forename><surname>Gurrin</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">U</forename><surname>Kruschwitz</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Caputo</surname></persName>
		</editor>
		<meeting><address><addrLine>Switzerland, Cham</addrLine></address></meeting>
		<imprint>
			<publisher>Springer Nature</publisher>
			<date type="published" when="2023">2023. 2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<monogr>
		<author>
			<persName><forename type="first">R</forename><surname>Alkhalifa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Yoong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Kochkina</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Zubiaga</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Liakata</surname></persName>
		</author>
		<idno>CoRR abs/2008.13160</idno>
		<ptr target="https://arxiv.org/abs/2008.13160.arXiv:2008.13160" />
		<title level="m">QMUL-SDS at checkthat! 2020: Determining COVID-19 tweet check-worthiness using an enhanced CT-BERT with numeric expressions</title>
				<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">The climate change twitter dataset</title>
		<author>
			<persName><forename type="first">D</forename><surname>Effrosynidis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">I</forename><surname>Karasakalidis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Sylaios</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Arampatzis</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.eswa.2022.117541</idno>
		<ptr target="https://doi.org/10.1016/j.eswa.2022.117541" />
	</analytic>
	<monogr>
		<title level="j">Expert Systems with Applications</title>
		<imprint>
			<biblScope unit="volume">204</biblScope>
			<biblScope unit="page">117541</biblScope>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<monogr>
		<author>
			<persName><forename type="first">R</forename><surname>Alkhalifa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><forename type="middle">M</forename><surname>Bilal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Borkakoty</surname></persName>
		</author>
		<author>
			<persName><surname>Romain</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Deveaud</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Luis</forename><surname>El-Ebshihy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Gabriela</forename><surname>Espinosa-Anke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Gonzalez-Saez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Galuscáková</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Goeuriot</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Kochkina</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Liakata</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Loureiro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Mulhem</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Piroi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Popel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><forename type="middle">T</forename><surname>Servan</surname></persName>
		</author>
		<author>
			<persName><surname>Madabushi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Zubiaga</forename><surname>Arkaitz</surname></persName>
		</author>
		<ptr target="https://api.semanticscholar.org/CorpusID:259953335" />
		<title level="m">Extended overview of the clef-2023 longeval lab on longitudinal evaluation of model performance</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b25">
	<analytic>
		<title level="a" type="main">Opinions are made to be changed: Temporally adaptive stance classification</title>
		<author>
			<persName><forename type="first">R</forename><surname>Alkhalifa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Kochkina</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Zubiaga</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2021 Workshop on Open Challenges in Online Social Networks</title>
				<meeting>the 2021 Workshop on Open Challenges in Online Social Networks</meeting>
		<imprint>
			<date type="published" when="2021">2021</date>
			<biblScope unit="page" from="27" to="32" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b26">
	<analytic>
		<title level="a" type="main">Bert: Pre-training of deep bidirectional transformers for language understanding</title>
		<author>
			<persName><forename type="first">J</forename><surname>Devlin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M.-W</forename><surname>Chang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Toutanova</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</title>
		<title level="s">Long and Short Papers</title>
		<meeting>the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</meeting>
		<imprint>
			<date type="published" when="2019">2019</date>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page" from="4171" to="4186" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b27">
	<monogr>
		<author>
			<persName><forename type="first">Y</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Ott</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Goyal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Du</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Joshi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Levy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Lewis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Zettlemoyer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Stoyanov</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1907.11692</idno>
		<title level="m">Roberta: A robustly optimized bert pretraining approach</title>
				<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b28">
	<monogr>
		<title level="m">Proceedings of Working Notes of CLEF 2024 -Conference and Labs of the Evaluation Forum, CEUR Workshop Proceedings</title>
				<editor>
			<persName><forename type="first">G</forename><surname>Faggioli</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Galuščáková</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><forename type="middle">G S</forename><surname>Herrera</surname></persName>
		</editor>
		<meeting>Working Notes of CLEF 2024 -Conference and Labs of the Evaluation Forum, CEUR Workshop Proceedings<address><addrLine>Aachen</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
