<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Overview of the CLEF 2024 SimpleText Task 2: Identify and Explain Difficult Concepts Notebook for the SimpleText Lab at CLEF 2024</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Giorgio</forename><forename type="middle">Maria</forename><surname>Di</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">University of Padova</orgName>
								<address>
									<settlement>Padova</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Federica</forename><surname>Vezzani</surname></persName>
							<email>federica.vezzani@unipd.it</email>
							<affiliation key="aff0">
								<orgName type="institution">University of Padova</orgName>
								<address>
									<settlement>Padova</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Vanessa</forename><surname>Bonato</surname></persName>
							<email>vanessa.bonato@unipd.it</email>
							<affiliation key="aff0">
								<orgName type="institution">University of Padova</orgName>
								<address>
									<settlement>Padova</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Hosein</forename><surname>Azarbonyad</surname></persName>
							<affiliation key="aff1">
								<orgName type="department">Elsevier</orgName>
								<address>
									<country key="NL">The Netherlands</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Jaap</forename><surname>Kamps</surname></persName>
							<affiliation key="aff2">
								<orgName type="institution">University of Amsterdam</orgName>
								<address>
									<settlement>Amsterdam</settlement>
									<country key="NL">The Netherlands</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Liana</forename><surname>Ermakova</surname></persName>
							<email>liana.ermakova@univ-brest.fr</email>
							<affiliation key="aff3">
								<orgName type="institution" key="instit1">Université de Bretagne Occidentale</orgName>
								<orgName type="institution" key="instit2">HCTI</orgName>
								<address>
									<settlement>Brest</settlement>
									<country key="FR">France</country>
								</address>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff4">
								<address>
									<settlement>Grenoble</settlement>
									<country key="FR">France</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Overview of the CLEF 2024 SimpleText Task 2: Identify and Explain Difficult Concepts Notebook for the SimpleText Lab at CLEF 2024</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">EAA02E0968A3EE71D147E31D15394B4F</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T17:56+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>automatic text simplification, terminology, background knowledge, scientific article, science popularization, contextualization, term difficulty (L. Ermakova) https://simpletext-project.com/ (L. Ermakova) 0000-0001-9709-6392 (G. Di Nunzio)</term>
					<term>0000-0001-9709-6392 (F. Vezzani)</term>
					<term>0009-0002-9918-282X (V. Bonato)</term>
					<term>0000-0002-6614-0087 (J. Kamps)</term>
					<term>0000-0002-7598-7474 (L. Ermakova)</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>In this paper, we present an overview of the "Task 2: Complexity Spotting, Identifying and explaining difficult concepts" within the context of the Automatic Simplification of Scientific Texts (SimpleText) lab, run as part of CLEF 2024. The primary objective of the SimpleText lab is to advance the accessibility of scientific information by facilitating automatic text simplification, thereby promoting a more inclusive approach to scientific knowledge dissemination. Task 2 focuses on complexity spotting within scientific texts (passage). Thus, the goal is to detect the terms/concepts that require specific background knowledge for understanding the passage, assess their complexity for non-experts, and provide explanations for these detected difficult concepts. A total of 39 submissions were received for this task, originating from 12 distinct teams. In this paper, we describe the data collection process, task configuration, and evaluation methodology employed. Additionally, we provide a brief summary of the various approaches adopted by the participating teams.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Despite digitalization making scientific literature more accessible to the public, a major barrier persists: the high complexity of scientific texts. Non-experts struggle to understand these texts due to a lack of background knowledge and specialized terminology. Even native speakers find it challenging to understand terms outside their expertise. Although those with basic education can somewhat understand popular science publications, scientific articles are mostly ignored by the citizens. Understanding terminology is crucial for comprehending scientific information. Comprehension of the term implies grasping the concept it represents without the need for an explicit definition. Definitions provide clear explanations of scientific terms, making complex ideas more understandable. Providing accurate definitions and background knowledge can reduce the risk of misinterpreting scientific information and help readers connect new information with what they already know, facilitating better integration and retention of new concepts.</p><p>While traditional simplification methods focus on removing complex terms and structures to improve readability <ref type="bibr" target="#b0">[1]</ref>, providing term definitions and background knowledge could help to make scientific texts more accessible, comprehensible, and meaningful to readers, enabling them to engage with the complex scientific information more effectively. Moreover, Scientific concepts often require precise terminology to avoid ambiguity and ensure clear communication among experts. These terms have specific meanings that cannot be easily replaced by simpler words without losing accuracy. Besides, many scientific ideas are inherently complex and cannot be adequately described with simple language.</p><p>Readers can recognize when they need definitions or clarifications for unfamiliar terms, reflecting their awareness of comprehension gaps. This awareness highlights their perception of difficulty with unfamiliar terminology. Thus, we argue that a text simplification method should provide essential information to help readers understand complex scientific concepts instead of deleting difficult terms. This objective is one of the crucial points of the CLEF 2024 SimpleText lab.</p><p>The CLEF 2024 SimpleText track <ref type="foot" target="#foot_0">1</ref> is an evaluation lab that follows up on the CLEF 2021 SimpleText Workshop <ref type="bibr" target="#b1">[2]</ref>, and CLEF 2022-2023 SimpleText Track <ref type="bibr" target="#b2">[3,</ref><ref type="bibr" target="#b3">4]</ref>.</p><p>The track offers valuable data and benchmarks to facilitate discussions on the challenges associated with automatic text simplification. The CLEF 2024 SimpleText track is based on four interrelated tasks:</p><p>1. Task 1 on Content Selection: retrieve passages to include in a simplified summary. 2. Task 2 on Complexity Spotting: identify and explain diffficult concepts. 3. Task 3 on Text Simplification: simplify scientific text. <ref type="bibr" target="#b3">4</ref>. Task 4 on SOTA?: track the state-of-the-art in scholarly publications. This paper focuses on the second task of complexity spotting. The goal of this task is to detect difficult terms and provide contextual explanations for them. Identifying and effectively explaining difficult terms is crucial for promoting accessibility and comprehension of scientific texts. Please refer for details of the other tasks to the overview papers of Task 1 <ref type="bibr" target="#b4">[5]</ref> and Task 3 <ref type="bibr" target="#b5">[6]</ref>, Task 4 <ref type="bibr" target="#b6">[7]</ref>, as well as the Track overview paper <ref type="bibr" target="#b7">[8]</ref>.</p><p>A total of 45 teams registered for our SimpleText track at CLEF 2024. A total of 20 teams submitted 207 runs in total for the Track, of which 13 teams submitted a total of 46 runs for Task 2. The statistics for the Task 2 runs submitted are presented in Table <ref type="table" target="#tab_0">1</ref>. However, some runs had problems that we could not resolve. We do not detail them in the paper as well as the 0-scored runs.</p><p>The rest of this paper is structured in the following way. A comprehensive description of the Task 2 is presented in Section 2. Following that, Section 3 provides an overview of the dataset used, including its composition, size, and relevant characteristics. In Section 4, the paper discusses the evaluation metrics employed to assess the performance of the participants' runs. Section 5 delves into the details of the systems and approaches employed by the participants. In Section 6, we discuss the results of the official submissions. We end with Section 7 discussing the results and findings, and lessons for the future.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Task description</head><p>The goal of this task is to identify key concepts that need to be contextualized with a definition, example or use case, and provide useful and understandable explanations for them. Thus, there are three subtasks:</p><p>• Task2.1: To predict what the terms are in a passage of a document and the difficulty of the concepts they designate (easy/medium/difficult). • Task 2.2: To generate a definition and an explanation for each difficult term. • Task 2.3: To retrieve the provided definitions of the difficult terms in "correct" order.</p><p>In Task 2.1, for each passage of a document, participants should provide a list of terms with corresponding scores (easy/medium/difficult) of the concepts they designate. Passages (sentences) are considered to be independent, that is term repetition is allowed (the same term can be detected in different sentences, even in the same document). Detected terms, their spans, and their difficulty will be evaluated. Both qualitative (manual review by terminologists) and quantitative metrics (recall and precision of the extracted terms) will be used to evaluate participants' results.</p><p>In Task 2.2, for each term that refers to a difficult concept (those that have been evaluated with the highest level of difficulty), participants should provide the definition and explanation which will be evaluated both from a qualitative point of view (manual review by terminologists) and from a quantitative point of view (overlapping text measures, for example, BLEU score <ref type="bibr" target="#b8">[9]</ref>).</p><p>In Task 2.3, participants should rank the set of definitions provided for the difficult terms in a way that the "best" definitions are ranked higher in the list of definitions. In particular, for each term there will be one manual definition (considered the best one) and two automatically generated good definitions that should be placed at the top of the list of retrieved definitions.. Quantitative metrics (for example, P@1, P@3, rank correlation measures) will be used to evaluate participants' results.</p><p>In general, we asked participants who wanted to run experiments on Task 2.2 to accomplish Task 2.1 first. On the other hand, Task 2.1 and Task 2.3 can be performed independently.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Data</head><p>The corpus of Task 2 is based on the sentences in high-ranked abstracts to the requests of Task 1. and collected in 2023 <ref type="bibr" target="#b3">[4]</ref>. A total of 175 documents and 1,077 sentences were used to generate the training and test data. In particular, we had 115 documents and 576 sentences for building the training set and 60 documents and 501 sentences for building the test set.</p><p>We provide a dataset for training the systems before the evaluation phase and a test set for the evaluation phase. In particular, the dataset comprises the following files:</p><p>• The documents and their sentences.</p><p>• Terms manually extracted from each sentence and their relative difficulty.</p><p>• Definitions and the explanations provided by the experts for the difficult terms.</p><p>• Definitions automatically generated by a large language model. For the training set, we engaged 21 experts to manually annotate each document, identifying the terms in each sentence, assessing their difficulty, and providing definitions and explanations for each difficult term. This effort resulted in the generation of 1,609 terms and 899 definitions and explanations.</p><p>To further analyze the consistency among experts, we deliberately assigned the same documents to multiple experts in some instances.</p><p>Additionally, for each term accompanied by a definition, we created two "good" definitions and two "bad" definitions. This was done to develop a set of definitions for ranking in Task 2.3, leading to a total of 2,356 sentences, evenly split between good and bad definitions.</p><p>Beyond this initial training set, we introduced an additional set of files produced by an external expert who reviewed the annotations of the 21 experts. This secondary set, referred to as the validation set, included the expert's additions of missing terms, definitions, or both. This review added 677 terms, 960 definitions, and 3,732 generated definitions (equally divided between good and bad) to the training data.</p><p>For the test set, we asked the external expert to annotate the remaining 60 documents. A total of 1,440 terms were extracted and 424 definitions were written from the 501 sentences of the test set. An additional 3,816 definitions (equally distributed between good and bad definitions) were also added.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.">Annotation Process</head><p>A first round of annotation was performed by a number of experts while a second round of validation was performed on the same set of documents by an external expert in order to look for additional (maybe missing) terms and definitions.</p><p>The process of annotation of the dataset consisted of two main phases:</p><p>1. the extraction of candidate terms from scientific abstracts, and 2. the construction of a collection of definitions of the concepts designated by candidate terms.</p><p>Concerning the first phase, we considered that terminology is a "set of designations and concepts belonging to one domain or subject" (ISO 1087: 2019 <ref type="bibr" target="#b9">[10]</ref>). The term, therefore, is a "designation that represents a general concept by linguistic means". By reading the abstracts, we identified the subjects or domains of knowledge that each abstract respectively deals with. Some examples of subjects or domains are the medical domain, drone technology and autonomous vehicle technology. Taking subjects and domains into consideration, we identified and extracted candidate terms in the texts of the abstracts. We refer to extracted terms as candidate terms due to the absence of validation of the result of term extraction on the part of experts.</p><p>The second phase concerned the construction of a collection of definitions of the concepts designated at the linguistic level by candidate terms. This phase involved two different stages: 1) the retrieval of definitions of the concepts, and 2) the transformation -where necessary -of source definitions into intensional definitions. In line with the ISO 1087: 2019, we adopt the view according to which an intensional definition is a "definition that conveys the intension of a concept by stating the immediate generic concept and the delimiting characteristic(s)".</p><p>Specifically, we retrieved definitions in different types of sources: general language dictionaries, specialized dictionaries, websites, papers and quotations included in websites or papers. In particular, the consultation of full-text articles from which abstracts were taken proved to be a useful method to retrieve definitions for specialized concepts. In many cases, the provided definition is a direct quotation of a definition contained in a source. In other cases, we adopted different approaches to the formulation of definitions. For example, we reformulated the source definitions or we embedded in a single definition information contained in more than one source.</p><p>In the first round of the annotation, the first set of experts were also asked to write explanations as a more natural and less structured way to clarify and make more intelligible a concept.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.">Input format</head><p>The training, validation, and test data are provided in TSV formats.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.1.">Documents and sentences</head><p>The dataset containing the documents and sentences will be stored in a file with the following format: </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.2.">Terms extracted</head><p>The terms manually extracted terms are stored in a file with the following information:</p><p>• snt_id: the identifier of each sentence </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.3.">Definitions and explanations</head><p>The definitions and explanations for the difficult terms are stored in a file with the following information:</p><p>• snt_id: the identifier of each sentence • term: the term extracted by the user • definition: the definition of the term • explanation: the explanation of the term • exp_id: the identifier of the expert who annotated that sentence (this column is not present in the validation files) </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.4.">Definitions generated</head><p>The definitions automatically generated are stored in a file with the following information:</p><p>• snt_id: the identifier of each annotated sentence • term: the term extracted by the expert • definition: the definition that has been used to generate the positive/negative definitions • positive: two definitions generated automatically that provide a "good" alternative to the "manual" definition. The two definitions are separated by a pipe "|" symbol. + • negative: two definitions generated automatically that provide a "wrong" alternative to the "manual" definition. The two definitions are separated by a pipe "|" symbol.+</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3.">Output format</head><p>Results should be provided in a TSV format or JSON format. A tabular example (TSV) of the output for Task 2.1 and Task 2.2 is shown in Figure <ref type="figure">1</ref>.</p><p>A JSON example of the same output is shown in Figure <ref type="figure">2</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3.2.">Task 2.3</head><p>For Task 2.3, the output file will be more similar to a TREC file. It must contain the following fields:</p><p>1. run_id: Run ID starting with &lt;team_id&gt;_&lt;task_id&gt;_&lt;method_used&gt;, e.g., UBO_Task2.3_TFIDF 2. manual: whether the run is manual {0 = no, 1 = yes} 3. snt_id: a unique passage (sentence) identifier from the input file of the test set 4. term: term for which definitions must be ranked 5. def_id: a unique identifier of the definition to be ranked 6. rank: an integer to specify the rank of this definition (1 highest rank, 2 second highest rank, . . . )</p><p>A tabular example (TSV) of the output for Task 2.3 is shown in Figure <ref type="figure">3</ref>.</p><p>A JSON example of the same output is shown in Figure <ref type="figure" target="#fig_2">4</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Evaluation metrics</head><p>In order to be as consistent as possible with the previous editions of SimpleText <ref type="bibr" target="#b3">[4]</ref>, we evaluated</p><p>• Task 2.1, difficult concept spotting, in terms of recall and precision; • Task 2.2, generation of definitions, in terms of BLEU score; • Task 2.3, the ranking of definitions, with precision@1 and precision@5. In particular, for Task 2.1, we wanted to evaluate both the recall and precision of all the terms, regardless of the difficulty of the concept, and of the difficult concepts only.</p><p>For Task 2.2, for the BLEU score we tried different ranges of values of the parameter n for the overlapping n-grams. <ref type="foot" target="#foot_1">2</ref>For Task 2.3, given the minimal number of runs submitted for Task 3 and the low significance of those results, no analysis will be presented in this paper.</p><p>In the future, a qualitative analysis will also be performed in order to study the problems of term identification and the generation of definitions. In addition, we will manually evaluate the provided explanations in terms of their usefulness with regard to a query as well as their complexity for a general </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Participant's Approaches</head><p>In this section, we describe the main approaches of each participant who submitted at least one run of a model to be evaluated on the test set. <ref type="foot" target="#foot_2">3</ref>AB&amp;DPV <ref type="bibr" target="#b10">[11]</ref> submitted one run, employing natural language processing techniques to identify difficult terms within passages. They generated definitions for these terms or retrieved them from sources like Wikipedia. However, they did not submit runs on the test set.</p><p>AIIRLab <ref type="bibr" target="#b11">[12]</ref> submitted three runs, utilizing LLaMA3 and Mistral language models. Their approach included prompt engineering and reinforcement learning with human feedback to enhance the quality of outputs generated by the LLaMA model. Dajana&amp;Kathy <ref type="bibr" target="#b12">[13]</ref> submitted one run, using the LLAMA-2 13B model. No further information about their approach was provided.</p><p>Frane&amp;Andrea <ref type="bibr" target="#b13">[14]</ref> submitted one run. No additional details about their methodology were given.</p><p>Sharingans <ref type="bibr" target="#b14">[15]</ref> submitted one run for, fine-tuning the GPT-3.5 turbo model for selecting difficult terms and generating definitions and explanations. They employed prompt-engineering techniques to create specific prompts that guided the model in producing accurate and contextually relevant definitions.</p><p>SINAI <ref type="bibr" target="#b15">[16]</ref> submitted three runs, applying learning cues without prior examples to the GPT-4-Turbo model. They used the OpenAI API in Python to interact with the model, facilitating the integration of GPT-4-Turbo into their workflow. team1_Petra_and_Regina <ref type="bibr" target="#b16">[17]</ref> submitted one run, combining named entity recognition (NER) techniques with rule-based approaches to identify and extract entities such as proteins, genes, and chemical compounds. They utilized spaCy for NER and developed custom rules for entity extraction.</p><p>Tomislav&amp;Rowan <ref type="bibr" target="#b17">[18]</ref> submitted two runs. They created prompts for the LLAMA-2 13B model to extract three scientific terms from each source sentence and then prompted the model to return a difficulty rating. Definitions for the difficult terms were retrieved from Wikipedia.</p><p>UAms <ref type="bibr" target="#b18">[19]</ref> submitted three runs. They employed an idf-based term weighting to identify the rarest terms for Task 2.1. For Task 2.3, they developed a method to rank definitions or explanations for given sentence-term pairs by examining the textual similarity of the provided sentences.</p><p>UBO <ref type="bibr" target="#b19">[20]</ref> submitted one run, using the Small Language Model, Phi3 mini, without fine-tuning, employing a one-shot prompt approach.</p><p>UNIPD <ref type="bibr" target="#b20">[21]</ref> submitted three runs, focusing on identifying and explaining difficult content using Large Language Models (LLMs) to enhance text simplification. They iteratively experimented with various prompting strategies to optimize model performance for this task.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Results</head><p>In this Section, we present the results on the test set for Task 2.1 and Task 2.2. At present time, the results for Task 2.3 are still ongoing (with only two runs by one participant) and will be made available in the future.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.1.">Test Results</head><p>The results on the test set are summarized in Table <ref type="table">?</ref>?. For each run, we report:</p><p>• the recall of all the terms, independently from the level of difficulty; • the precision of all the terms, independently from the level of difficulty; • the F1 score of all the terms, independently from the level of difficulty; • the recall of the difficult terms; • the precision of the difficult terms; • the F1 score of the difficult terms; • the BLEU score computed for bigrams (ngrams from n = 1 to n = 4). 4   For recall, precision, and F1 we report both the "overall" scores computed by summing up all the data for all the retrieved and predicted terms for all the sentences, and the "average" scores computed by averaging the recall and precision of the predicted terms per sentence.</p><p>In Table <ref type="table">2</ref>, we report the results for all the terms; in Table <ref type="table">3</ref>, we show the results for the difficult terms only; in Table <ref type="table">4</ref>, we present the results of the BLEU scores. recall and precision are sufficiently good but suboptimal when compared to the state-of-the-art models (of course, we need to take into consideration that this is the first time participants dealt with this new dataset).</p><p>Our main findings are the following. First, the runs submitted by the participants to this task are quite stable in terms of recall-precision performances when dealing with all the terms or the difficult ones. Independently from the difficulty of terms, the models proposed by the participants can achieve precision higher than .50 across a range of recall values. The best runs can achieve an average recall in the range of 0.3 -0.5 while obtaining a precision between 0.4 and 0.7. It is interesting to see that the best average precision (0.7604) for all the terms is achieved by the experiment that performed a manual cleaning and intervention of the output of ChatGPT. If we focus only on the difficult terms, the recall in general decreases while the precision increases, which means that a smaller amount of difficult terms are found but the system is very precise in detecting them. Second, the BLEU score of the generated definitions is also relatively stable ranging from 0.2 to 0.3 for n = 1 and and from 0.1 to 0.2 for n = 2 for any recall values. Third, the best performing runs are usually those that has some analysis of the optimal prompting or a manual interaction with the model. This is inline with the latest research studies on this issue <ref type="bibr" target="#b22">[23]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.3.">Qualitative Analysis</head><p>In this section, we present a qualitative analysis in which we comparatively evaluate the definitions of concepts provided by an expert in terminology and the definitions proposed by different participants. In the following analysis, the first definition is the one provided by the terminologist and the second one is the definition provided by the participants. For each of the definitions submitted by the participants of task 2.2, the BLEU score is computed. This score is calculated by computing the degree of similarity that is established between the concept definition provided by and the definition formulated by the participants. For each analyzed participant, we analyze form a qualitative perspective the definitions that respectively obtained the highest and the loweest BLEU score computed with n = 2.</p><p>For what concerns the concept designated by the term "MEC server", we analyze the following two definitions. The definition provided by the participant is the one that obtained the highest BLEU score which amounts to 0.3223. The term is included in the sentence identified by the code G06.2_2895666646_7.</p><p>• Server of Multi-access edge computing, which is a cloud service running at the edge of a network and performing specific tasks that would otherwise be processed in centralized core or cloud infrastructures. • A Multi-Access Edge Computer (MEC) server is a type of server technology located at the edge of a communication network to reduce the latency and increase the speed of data delivery.</p><p>From a qualitative viewpoint, it is possible to observe that both definitions include the indication of the extended form of the acronym "MEC" and refer to the location of the server with respect to a network. However, some terms that are present in the expert's definition do not match the terms included in the participant's definition, as in the case of "cloud service", "task", "centralized core infrastructure" and "cloud infrastructure". The definition provided by the same participant that obtained the lowest BLEU score (0.0247) is the definition of the concept designated by the term "NCI". The sentence in which the term is present is identified by the code G08.2_1607424157_4.</p><p>• Measure of collective behaviour based on financial news on the Web, which captures the average mutual similarity between the documents and entities in the financial corpus. • NCI stands for Named Complexity Index. As it can be observed, the text provided by the participant aims at explaining the abbreviated form "NCI" by indicating the extended version of the acronym rather than defining the concept designated by the term itself. Indeed, the text does not express the intension (that is the main characteristics) of the concept and, as a consequence, it cannot be considered as a terminological definition.</p><p>Another participant, proposed the definition of the concept designated by the term "Windows 2000 machine" obtaining the highest BLEU score, amounting to 0.5939. The term is contained in the sentence whose identifier is G01.1_1522515958_7.</p><p>• Computer system that uses the operating system called Windows 2000.</p><p>• A computer that uses the Windows 2000 operating system.</p><p>The major difference between the two definitions lies in the indication of the generic concept, which is respectively identified as "computer system" by the terminologist, and as "computer" by the participant. In this case, "computer system" refers to the entire assembly of hardware and software (including the operating system) is the most accurate choice rather than "computer" which usually refers to the hardware only.</p><p>The definition provided by the same participant of the concept designated by the term "Mixed Logit model" obtained the lowest BLEU score (0.0023). The term is included in the sentence identified by the code G06.2_2982382045_4.</p><p>• Statistical model, that is fully general, for examining discrete choices, which allows for random taste variation across choosers, unrestricted substitution patterns across choices, and correlation in unobserved factors over time. • Model used in statistics. In the definition provided by the participant, an absence of delimiting characteristics of the concepts can be observed. As a matter of fact, the definition would not allow to distinguish the Mixed Logit model from other models used in the domain of statistics. In this sense, the text could be better classified as a generic explanation rather than a terminological definition.</p><p>The same procedure is applied to the definitions provided by another participant. The definition presented by the participant that reached the highest BLEU score (0.5298) refers to the concept designated by the term "GTAV". The sentence in which the term is included is identified by the code G06.2_2890116921_6.</p><p>• Action-adventure game developed by Rockstar North and published by Rockstar Games.</p><p>• Grand Theft Auto V, an action-adventure video game developed by Rockstar North.</p><p>As it can be observed, the definition presented by the participant includes elements that are not present in the definition provided by the expert, that are: 1) the extended form of the term, and 2) the specification concerning the fact that Grand Theft Auto V is a video game. The definition provided by the terminologist, however, also indicates the publisher of the game. In the definition provided by the participant, the generic concept is not placed as the first element of the text, thus not following the structure attributed to intensional definitions. The definition included by the participant that reached the lowest BLEU score (0.03556) is related to the concept designated by the term "Autoware". The term can be individuated in the sentence identified by the code G06.2_2931522054_8.</p><p>• Software stack platform for self-driving that is ROS-based, composed of an abundant set of selfdriving modules, such as sensing, localization, detection, planning, and actuation, and libraries that render it possible to operate and simulate autonomous vehicle. • An open-source software stack designed for self-driving vehicles. The difference between the two definitions consists in the richness of information that characterizes the definition of the expert, with respect to the amount of information provided in the second one. Moreover, the second definition does not follow the structure of intensional definitions.</p><p>We additionally perform the analysis of the definitions inserted by another participant. The definition provided by the participant of the concept designated by the term "Denial-of-Service" obtained the highest BLEU score (0.8678). The term is found in the sentence whose identifier is G06.2_2548923997_6.</p><p>• Cyber-attack in which the perpetrator seeks to make a machine or network resource unavailable to its intended users by temporarily or indefinitely disrupting services of a host connected to a network. • A cyber-attack where the perpetrator seeks to make a machine or network resource unavailable to its intended users by temporarily or indefinitely disrupting services of a host connected to the Internet.</p><p>In this circumstance, it is possible to observe that the two definitions present a high level of similarity.</p><p>The major difference lies in the respective usage of the term "network" and "Internet" to designate the connection used by the host in this type of cyber-attack. The participant also provided a definition of the concept designated by the term "Autoware", that presented the lowest BLEU score (0.0231) with respect to the expert's definition. The term can be found in the sentence identified as G06.2_2931522054_7.</p><p>• Software stack platform for self-driving that is ROS-based, composed of an abundant set of selfdriving modules, such as sensing, localization, detection, planning, and actuation, and libraries that render it possible to operate and simulate autonomous vehicle. • An open-source software platform designed to enable autonomy for various types of vehicles.</p><p>In the first definition, a greater amount of information is given. Another difference that can be observed is that the second definition. Finally, we evaluate the performance in terms of BLEU score obtained by two definitions inserted by another participant. We begin with the analysis of the definition that obtained the highest BLEU score (0.5621), related to the concept designated by the term "Windows 2000 machine". The term is comprised in a sentence whose identifier is G01.1_1522515958_7.</p><p>• Computer system that uses the operating system called Windows 2000.</p><p>• Computer system running the Windows 2000 operating system.</p><p>Both definitions are characterized by the presence of the same generic concept, linguistically designated by the term "computer system". The differences between the two definitions are: 1) the use of different verbs, and 2) the diverse structural arrangement of information. The same participant also provided a definition of the concept designated by the term "Dr Who CSR engine". This definition obtained the lowest BLEU score, amounting to 0.0240. The term is included in a sentence identified by the code G01.1_1522515958_8.</p><p>• Engine for continuous speech recognition developed in the context of the Microsoft research project Dr Who, which uses a unified language model that takes advantage of rule-based and data-driven approach. • Speech recognition engine used in MiPad. Also in this case, it can be observed that the amount of information provided in the context of the two definitions does not match.</p><p>To conclude, we propose a qualitative analysis concerning the results stemming from the task of term extraction performed by an expert in terminology and by different participants. In particular, we focus on the detection of lexical units classified as terms relevant for the domain by participants, that are however excluded from the list of terms extracted by the expert. From the observation of the sentence identified by the code G01.1_130055196_1, we noticed that a participant considered "PDA (Personal Digital Assistant)" as a term. This string of characters, however, does not correspond to a term. As a matter of fact, it is possible to identify two different terms designating the same concept: 1) "PDA", and 2) "Personal Digital Assistant". In particular, "PDA" is the acronym for "Personal Digital Assistant". Moreover, with specific reference to the sentence whose identifier is G01.1_135571562_4, the participant also selected "with a barcode reader" as a term. The correct term, however, is "barcode reader". Another string of characters that does not constitute a term is "regarded as important", contained in the sentence coded as G01.1_135571562_7. This is due to the fact that "regarded as important" does not designate a concept in specialized domains of knowledge. The lexical units "regarded" and "important" were also extracted as terms by another participant, in the context of the sentence coded as G01.1_135571562_7.</p><p>Nevertheless, even when considered as two different lexical units, they do not constitute designations of concepts in specialized fields of knowledge. Moreover, a participant considered cardinal numbers as constituting elements of terms, such as in "eight sets" and "five repetitions", both contained in the sentence coded as M1_13_1. These strings of characters should not be extracted as terms, due to the consideration that a correct term extraction would result in the extraction of the terms "set" and "repetition". Furthermore, the string of characters "clinicianu0027s personal assistant", included in the sentence identified by the code G01.1_1462481249_3, was also detected as a term by a participant. However, in this case two different terms can be individuated: 1) "clinician" and 2) "personal assistant". The string of characters "u0027" should not be considered as a constituting element of a term, as it represents the Unicode number for the Unicode character "'". The term "closest facilities", contained in the sentence whose identifier is G01.1_1000902583_3, was also regarded as a term by a participant. However, "closest" constitutes a superlative adjective. Considering that, the term that should be extracted is "facility".</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="7.">Conclusion and future work</head><p>The results of Task 2 of the CLEF 2024 SimpleText challenge have demonstrated the potential and limitations of current natural language processing (NLP) models in identifying and defining difficult complex within scientific texts, and ranking available definitions for those concepts. The task was divided into three subtasks: identifying terms and their difficulty (Task 2.1), generating definitions and explanations for difficult terms (Task 2.2), and ranking definitions (Task 2.3). The diversity of approaches taken by the participants showcased various strategies and methodologies in tackling these problems.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="7.1.">Summary of Findings</head><p>In Task 2.1, precision and recall metrics highlighted that while some systems could accurately identify terms, there was a general challenge in consistently predicting the difficulty level. Several approaches, using models like LLaMA and Mistral, showed promising results in term identification, but the precision for difficult terms varied significantly.</p><p>Task 2.2 focused on generating definitions, where the BLEU score was used to evaluate the overlap between generated and reference definitions. Here, the performance varied, with some models generating coherent definitions, while others struggled with accuracy and relevance. Some fine-tuned LLM model achieved notable BLEU scores, indicating effective use of reinforcement learning and prompt engineering.</p><p>For Task 2.3, which required ranking definitions, there was limited participation, and the results are still ongoing. The preliminary findings suggest that ranking automatically generated definitions in the correct order remains a significant challenge, and further analysis is needed to draw concrete conclusions.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="7.2.">Future Work</head><p>The task results point to several directions for future research and development. Future work should focus on improving the precision and recall of term identification, particularly for difficult terms. This may involve integrating more sophisticated context-aware models and leveraging domain-specific knowledge bases. The quality of generated definitions needs enhancement. Research into more advanced language models, fine-tuning techniques, and hybrid approaches combining rule-based and machine learning methods could yield better results. The current evaluation metrics provide a good starting point, but future tasks could benefit from more nuanced and context-sensitive metrics that better capture the quality and relevance of generated definitions and explanations. Incorporating human feedback iteratively in the model training process can significantly improve the quality of outputs.</p><p>In conclusion, the CLEF 2024 SimpleText challenge has provided valuable insights into the capabilities of current NLP models in understanding and processing complex scientific texts. Continued research and collaboration within the community will be essential in addressing the identified challenges and advancing the state of the art in this field.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>•</head><label></label><figDesc>doc_id: the identifier of the document • snt_id: the identifier of sentence • snt_source: the text of the sentence For example The users in an initial study [...] 2093013061 G10.1_2093013061_1 In this paper, we present an [...] 2093013061 G10.1_2093013061_2 From the data of an onboard [...]</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 2 : 2 Figure 3 :</head><label>223</label><figDesc>Figure 2: JSON Example for Task 2.1 and Task 2.2</figDesc><graphic coords="7,128.41,65.60,338.48,404.99" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Figure 4 :</head><label>4</label><figDesc>Figure 4: JSON Example for Task 2.3</figDesc><graphic coords="8,207.38,65.61,180.51,342.92" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1</head><label>1</label><figDesc>CLEF 2024 Simpletext Task 2 official run submission statistics</figDesc><table><row><cell>Task</cell><cell>AIIR Lab</cell><cell>AMATU</cell><cell>Arampatzis</cell><cell>Elsevier</cell><cell>L3S</cell><cell>LIA</cell><cell>PiTheory</cell><cell>Sharigans</cell><cell>SINAI</cell><cell>SONAR</cell><cell>AB/DPV</cell><cell>Dajana/Katya</cell><cell>Frane/Andrea</cell><cell>Petra/Regina</cell><cell>Ruby</cell><cell>Tomislav/Rowan</cell><cell>UAmsterdam</cell><cell>UBO</cell><cell>UniPD</cell><cell>UZH Pandas</cell><cell>Total</cell></row><row><cell>2.1</cell><cell>3</cell><cell></cell><cell>5</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell>1</cell><cell>3</cell><cell></cell><cell>1</cell><cell>1</cell><cell>1</cell><cell>1</cell><cell>1</cell><cell cols="4">2 1 1 3</cell><cell></cell><cell>24</cell></row><row><cell>2.2</cell><cell>3</cell><cell></cell><cell>5</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell>1</cell><cell>3</cell><cell></cell><cell>1</cell><cell></cell><cell>1</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell cols="2">1 3</cell><cell></cell><cell>18</cell></row><row><cell>2.3</cell><cell></cell><cell></cell><cell>2</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell>2</cell><cell></cell><cell></cell><cell></cell><cell>4</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head></head><label></label><figDesc>• term: the term extracted by the user • difficulty: the difficulty assigned by the expert ([e]asy/[m]edium/[d]ifficult)• exp_id: the identifier of the expert who annotated that sentence (this column is not present in the validation files)</figDesc><table><row><cell>snt_id</cell><cell>term</cell><cell>difficulty</cell><cell>exp_id</cell></row><row><cell>G06.2_2968176166_5</cell><cell>automated</cell><cell>e</cell><cell>1</cell></row><row><cell>G06.2_2968176166_5</cell><cell>bayesian</cell><cell>d</cell><cell>1</cell></row><row><cell>G01.1_1019677957_1</cell><cell>mobile technology</cell><cell>m</cell><cell>2</cell></row><row><cell>G01.1_1019677957_1</cell><cell>mobile emerging carrier</cell><cell>d</cell><cell>2</cell></row><row><cell>G01.1_1019677957_1</cell><cell>personal digital assistant</cell><cell>d</cell><cell>2</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">https://simpletext-project.com</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_1">https://cran.r-project.org/web/packages/sacRebleu/vignettes/sacReBLEU.html</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_2">Some participant decided to submit the experimental result on the training set which will be useful for future post-hoc analyses.</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgments</head><p>This work is partially supported by the HEREDITARY Project, as part of the European Union's Horizon Europe research and innovation programme under grant agreement No GA 101137074.</p><p>This research was funded, in whole or in part, by the French National Research Agency (ANR) under the project ANR-22-CE23-0019-01.</p><p>We would like to thank Jaap Kamps, Valentin Laimé, Radia Hannachi, Silvia Araújo, Pierre De Loor, Olga Popova, Diana Nurbakova, Quentin Dubreuil, and all the other colleagues and participants who helped run this track.</p></div>
			</div>

			<div type="annex">
<div xmlns="http://www.tei-c.org/ns/1.0"><p>In, Figure <ref type="figure">5</ref> and Figure <ref type="figure">6</ref> the precision and recall results for the overall and averaged measures. We      </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.2.">Quantitative Analysis</head><p>The results shown in the previous section reveal that the use of large language models for the extraction of terms, the assessment of the difficulty of these terms, and the generation of the definitions to explain the difficult concepts are at an initial stage that will open new perspective in the Automatic Term Extraction panorama. In particular, compared to the recent results and surveys (see <ref type="bibr" target="#b21">[22]</ref>), the values of</p></div>			</div>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">A Word-Complexity Lexicon and A Neural Readability Ranking Model for Lexical Simplification</title>
		<author>
			<persName><forename type="first">M</forename><surname>Maddela</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Xu</surname></persName>
		</author>
		<ptr target="https://www.aclweb.org/anthology/D18-1410" />
	</analytic>
	<monogr>
		<title level="m">Proc. of EMNLP 2018, ACL</title>
				<meeting>of EMNLP 2018, ACL<address><addrLine>Brussels, Belgium</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page" from="3749" to="3760" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Overview of simpletext 2021 -CLEF workshop on text simplification for scientific information access</title>
		<author>
			<persName><forename type="first">L</forename><surname>Ermakova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Bellot</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Braslavski</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Kamps</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Mothe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Nurbakova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Ovchinnikova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Sanjuan</surname></persName>
		</author>
		<idno type="DOI">10.1007/978-3-030-85251-1_27</idno>
		<idno>doi:</idno>
		<ptr target="10.1007/978-3-030-85251-1\_27" />
	</analytic>
	<monogr>
		<title level="m">Experimental IR Meets Multilinguality, Multimodality, and Interaction -12th International Conference of the CLEF Association, CLEF 2021, Virtual Event</title>
		<title level="s">Lecture Notes in Computer Science</title>
		<editor>
			<persName><forename type="first">K</forename><forename type="middle">S</forename><surname>Candan</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">B</forename><surname>Ionescu</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">L</forename><surname>Goeuriot</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">B</forename><surname>Larsen</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">H</forename><surname>Müller</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Joly</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><surname>Maistro</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">F</forename><surname>Piroi</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><surname>Faggioli</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</editor>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2021">September 21-24, 2021. 2021</date>
			<biblScope unit="volume">12880</biblScope>
			<biblScope unit="page" from="432" to="449" />
		</imprint>
	</monogr>
	<note>Proceedings</note>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Overview of the CLEF 2022 simpletext lab: Automatic simplification of scientific texts</title>
		<author>
			<persName><forename type="first">L</forename><surname>Ermakova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Sanjuan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Kamps</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Huet</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Ovchinnikova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Nurbakova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Araújo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Hannachi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">É</forename><surname>Mathurin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Bellot</surname></persName>
		</author>
		<idno type="DOI">10.1007/978-3-031-13643-6_28</idno>
		<idno>doi:</idno>
		<ptr target="10.1007/978-3-031-13643-6\_28" />
	</analytic>
	<monogr>
		<title level="m">Experimental IR Meets Multilinguality, Multimodality, and Interaction -13th International Conference of the CLEF Association, CLEF 2022</title>
		<title level="s">Lecture Notes in Computer Science</title>
		<editor>
			<persName><forename type="first">A</forename><surname>Barrón-Cedeño</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><forename type="middle">D S</forename><surname>Martino</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><forename type="middle">D</forename><surname>Esposti</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">F</forename><surname>Sebastiani</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">C</forename><surname>Macdonald</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><surname>Pasi</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Hanbury</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><surname>Potthast</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><surname>Faggioli</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</editor>
		<meeting><address><addrLine>Bologna, Italy</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2022">September 5-8, 2022. 2022</date>
			<biblScope unit="volume">13390</biblScope>
			<biblScope unit="page" from="470" to="494" />
		</imprint>
	</monogr>
	<note>Proceedings</note>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Overview of the CLEF 2023 simpletext lab: Automatic simplification of scientific texts</title>
		<author>
			<persName><forename type="first">L</forename><surname>Ermakova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Sanjuan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Huet</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Azarbonyad</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Augereau</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Kamps</surname></persName>
		</author>
		<idno type="DOI">10.1007/978-3-031-42448-9_30</idno>
		<ptr target="https://doi.org/10.1007/978-3-031-42448-9_30.doi:10.1007/978-3-031-42448-9\_30" />
	</analytic>
	<monogr>
		<title level="m">Experimental IR Meets Multilinguality, Multimodality, and Interaction -14th International Conference of the CLEF Association, CLEF 2023</title>
		<title level="s">Lecture Notes in Computer Science</title>
		<editor>
			<persName><forename type="first">A</forename><surname>Arampatzis</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">E</forename><surname>Kanoulas</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">T</forename><surname>Tsikrika</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S</forename><surname>Vrochidis</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Giachanou</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">D</forename><surname>Li</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><surname>Aliannejadi</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><surname>Vlachos</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><surname>Faggioli</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</editor>
		<meeting><address><addrLine>Thessaloniki, Greece</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2023">September 18-21, 2023. 2023</date>
			<biblScope unit="volume">14163</biblScope>
			<biblScope unit="page" from="482" to="506" />
		</imprint>
	</monogr>
	<note>Proceedings</note>
</biblStruct>

<biblStruct xml:id="b4">
	<monogr>
		<title level="m" type="main">Overview of the CLEF 2024 SimpleText task 1: Retrieve passages to include in a simplified summary</title>
		<author>
			<persName><forename type="first">E</forename><surname>Sanjuan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Huet</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Kamps</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Ermakova</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<monogr>
		<author>
			<persName><forename type="first">L</forename><surname>Ermakova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Laimé</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Mccombie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Kamps</surname></persName>
		</author>
		<title level="m">Overview of the CLEF 2024 SimpleText task 3: Simplify scientific text</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<monogr>
		<author>
			<persName><forename type="first">J</forename><surname>Souza</surname></persName>
		</author>
		<title level="m">Overview of the CLEF 2024 SimpleText task 4: Track the state-of-the-art in scholarly publications</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Overview of the CLEF 2024 SimpleText track: Improving access to scientific texts for everyone</title>
		<author>
			<persName><forename type="first">L</forename><surname>Ermakova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Sanjuan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Huet</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Azarbonyad</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">M</forename><surname>Di Nunzio</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Vezzani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Souza</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Kamps</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Experimental IR Meets Multilinguality, Multimodality, and Interaction. Proceedings of the Fifteenth International Conference of the CLEF Association (CLEF 2024)</title>
		<title level="s">Lecture Notes in Computer Science</title>
		<editor>
			<persName><forename type="first">L</forename><surname>Goeuriot</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><forename type="middle">Q</forename><surname>Philippe Mulhem</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">D</forename><surname>Schwab</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">L</forename><surname>Soulier</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><forename type="middle">M D</forename><surname>Nunzio</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Galuščáková</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><forename type="middle">G S</forename><surname>De Herrera</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><surname>Faggioli</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</editor>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Bleu: a method for automatic evaluation of machine translation</title>
		<author>
			<persName><forename type="first">K</forename><surname>Papineni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Roukos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Ward</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W.-J</forename><surname>Zhu</surname></persName>
		</author>
		<idno type="DOI">10.3115/1073083.1073135</idno>
		<idno>doi:10. 3115/1073083.1073135</idno>
		<ptr target="https://doi.org/10.3115/1073083.1073135" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 40th Annual Meeting on Association for Computational Linguistics, ACL &apos;02, Association for Computational Linguistics</title>
				<meeting>the 40th Annual Meeting on Association for Computational Linguistics, ACL &apos;02, Association for Computational Linguistics<address><addrLine>USA</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2002">2002</date>
			<biblScope unit="page" from="311" to="318" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<monogr>
		<title level="m">Terminology work and terminology science -Vocabulary, Standard, International Organization for Standardization</title>
				<meeting><address><addrLine>Geneva, CH</addrLine></address></meeting>
		<imprint>
			<date type="published" when="1087">ISO1087. 2019. 2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">SimpleText</title>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">P</forename><surname>Varadi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Bartulović</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Scientific Text Made Simpler Through the Use of AI</title>
				<imprint>
			<date type="published" when="2024">2024. 2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<monogr>
		<title level="m" type="main">AIIR Lab Systems for CLEF 2024 SimpleText: Large Language Models for Text Simplification</title>
		<author>
			<persName><forename type="first">N</forename><surname>Largey</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Maarefdoust</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Durgin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Mansouri</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<monogr>
		<title level="m" type="main">Simplify It Like It&apos;s Hot: Making Complex Texts Easy to Digest</title>
		<author>
			<persName><forename type="first">K</forename><surname>Seng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Simunovic</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<monogr>
		<title level="m" type="main">Simplify It Like It&apos;s Hot: Making Complex Texts Easy to Digest</title>
		<author>
			<persName><forename type="first">A</forename><surname>Zečević</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Doljanin</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<monogr>
		<title level="m" type="main">Improving Scientific Text Comprehension: A Multi-Task Approach with GPT-3.5 Turbo and Neural Ranking</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">M</forename><surname>Ali</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Sajid</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Aijaz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Waheed</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Alvi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Samad</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Zero-shot Prompting on GPT-4-Turbo for Lexical Complexity Prediction</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">A</forename><surname>Ortiz-Zambrano</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Espin-Riofrio</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Montejo-Ráez</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">SINAI Participation in SimpleText Task 2 at CLEF 2024</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<monogr>
		<title level="m" type="main">AI Contributions to Simplifying Scientific Discourse in SimpleText</title>
		<author>
			<persName><forename type="first">R</forename><surname>Elagina</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Vučić</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2024">2024. 2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<monogr>
		<author>
			<persName><forename type="first">R</forename><surname>Mann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Mikulandric</surname></persName>
		</author>
		<title level="m">CLEF 2024 SimpleText Tasks 1-3: Use of LLaMA-2 for text simplification</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<monogr>
		<author>
			<persName><forename type="first">J</forename><surname>Bakker</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Yüksel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Kamps</surname></persName>
		</author>
		<title level="m">University of Amsterdam at the CLEF 2024 SimpleText Track</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<monogr>
		<title level="m" type="main">UBO NLP report on the SimpleText track at CLEF</title>
		<author>
			<persName><forename type="first">B</forename><surname>Vendeville</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Ermakova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">De</forename><surname>Loor</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2024">2024. 2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<monogr>
		<title level="m" type="main">UNIPD@SimpleText2024: A Semi-Manual Approach on Prompting ChatGPT for Extracting Terms and Write Terminological Definitions</title>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">M</forename><surname>Di Nunzio</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Vezzani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Gallina</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">A systematic review of Automatic Term Extraction: What happened in 2022?</title>
		<author>
			<persName><forename type="first">G</forename><surname>Di Nunzio</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Marchesin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Silvello</surname></persName>
		</author>
		<idno type="DOI">10.1093/llc/fqad030</idno>
		<ptr target="https://doi.org/10.1093/llc/fqad030.doi:10.1093/llc/fqad030" />
	</analytic>
	<monogr>
		<title level="j">Digital Scholarship in the Humanities</title>
		<imprint>
			<biblScope unit="volume">38</biblScope>
			<biblScope unit="page" from="41" to="47" />
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">How to write effective prompts for large language models</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Lin</surname></persName>
		</author>
		<idno type="DOI">10.1038/s41562-024-01847-2</idno>
		<ptr target="https://www.nature.com/articles/s41562-024-01847-2.doi:10.1038/s41562-024-01847-2,pub-lisher:NaturePublishingGroup" />
	</analytic>
	<monogr>
		<title level="j">Nature Human Behaviour</title>
		<imprint>
			<biblScope unit="volume">8</biblScope>
			<biblScope unit="page" from="611" to="615" />
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<ptr target=".org" />
	</analytic>
	<monogr>
		<title level="m">Working Notes of CLEF 2024: Conference and Labs of the Evaluation Forum</title>
		<title level="s">CEUR Workshop Proceedings</title>
		<editor>
			<persName><forename type="first">G</forename><surname>Faggioli</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Galuščáková</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><forename type="middle">G S</forename><surname>Herrera</surname></persName>
		</editor>
		<imprint>
			<publisher>CEUR-WS</publisher>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
