<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">University of Amsterdam at the CLEF 2022 SimpleText Track</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Femke</forename><surname>Mostert</surname></persName>
						</author>
						<author>
							<persName><forename type="first">Ashmita</forename><surname>Sampatsing</surname></persName>
						</author>
						<author>
							<persName><forename type="first">Mink</forename><surname>Spronk</surname></persName>
						</author>
						<author>
							<persName><forename type="first">David</forename><surname>Rau</surname></persName>
						</author>
						<author role="corresp">
							<persName><forename type="first">Jaap</forename><surname>Kamps</surname></persName>
							<email>kamps@uva.nl</email>
						</author>
						<author>
							<affiliation key="aff0">
								<orgName type="institution">University of Amsterdam</orgName>
								<address>
									<settlement>Amsterdam</settlement>
									<country key="NL">The Netherlands</country>
								</address>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff1">
								<orgName type="department">Evaluation Forum</orgName>
								<address>
									<addrLine>September 5-8</addrLine>
									<postCode>2022</postCode>
									<settlement>Bologna</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">University of Amsterdam at the CLEF 2022 SimpleText Track</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">CE829C4465B564F63F273AC2332458B6</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T03:23+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Information Storage and Retrieval</term>
					<term>Natural Language Processing</term>
					<term>Scientific Information Access</term>
					<term>Text Simplification</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>This paper reports on the University of Amsterdam's participation in the CLEF 2022 SimpleText track. The overall goal of removing barriers that prevent the general public from accessing scientific literature is of great importance to help users make sense of a world of misinformation and shallow opinions. We perform preliminary studies within the track's setup, analyzing the text complexity of searching a large set of academic abstracts in the context of popular science topics emerging in the news, with a specific focus at the relation between the topical relevance and the text complexity of the retrieved information. Our main findings are the following. First, we analyzed a large corpus of scientific abstracts and confirmed that these are highly complex on average, but that the variation is large and many abstracts with accessible readability levels exist. Second, we ran retrieval experiments and found that standard search ignores readability, yet filtering on the desirable reading level still retains competitive performance while avoiding retrieving relevant but incomprehensible results. Third, we ran complexity spotting experiments and found that straightforward lexical complexity or term frequency measures are strong indicators, but have to be combined with the importance of the concept in the broader context of the information request. Fourth, we ran a GPT-2 based text simplification model in a zero-shot way, resulting in conservative rewriting of abstracts, able to significantly reduce the text complexity. More generally, our results demonstrate that text complexity is an essential aspect to consider for improving non-expert access to scientific information, and opens up new routes to develop effective scientific information access technology tailored to needs of the general public.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>The advent of the internet and social media has been revolutionary in changing every aspect of information creation and information consumption. In the early years, many have lauded all the positives resulting from this, such as breaking down traditional barriers in access to information, as well as providing the means to publish anything by anyone <ref type="bibr" target="#b0">[1]</ref>. One of the prime examples of this is Wikipedia. This is leading in principle to ultimate democratization by giving every global citizen a voice, and ultimate equality where quality of arguments rather than the provenance of the author decides the outcome.</p><p>In recent years, the emphasis has shifted to the negatives resulting from this, as not only individual citizens have massively joined this open sphere, but also many other commercial or political actors have discovered its potential to influence citizens to suit their financial or political incentives, even leading to social and societal risks. Hence, one of the greatest challenges of today is how users can navigate in a world of misinformation and disinformation. This is a very hard and complex problem, and rather than believing in fairy tales and some magic wand that will make it disappear, we focus on a well-established antidote: objective, scientific information in the academic literature and associated data. Every citizen agrees on the importance of objective scientific evidence, yet at the same time they predominantly rely on shallow secondary information on the web and in social media. One of the main reasons for not accessing scientific information directly is that they presume the scientific literature is too difficult. The CLEF 2022 SimpleText track investigates the barriers that ordinary citizens face when accessing scientific literature head-on, by making available corpora and tasks to address different aspects of the problem. For details on the exact track setup, we refer to the Track Overview paper CLEF 2022 LNCS proceedings <ref type="bibr" target="#b1">[2]</ref>, as well as the detailed task overviews in the CEUR proceedings <ref type="bibr" target="#b2">[3,</ref><ref type="bibr" target="#b3">4,</ref><ref type="bibr" target="#b4">5]</ref>.</p><p>We conduct an extensive analysis of the corpus of scientific abstracts and the three tasks of the track: Task 1 on content selection and avoiding complexity; Task 2 on complexity spotting in extracted sentences from scientific abstracts; and Task 3 on text simplification proper rewriting sentences from these abstracts. The rest of this paper is structured as follows. Next, in Section 2 we discuss our experimental setup and the specific runs submitted. Section 3 discusses the results of our runs and provides a detailed analysis of the corpus and results for each task. We end in Section 4 by discussing our results and outlining the lesson learned.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Experimental Setup</head><p>In this section, we will detail our approach for the three CLEF 2022 SimpleText track tasks.</p><p>For details of the exact task setup and results we refer the reader to the detailed overview of the track in Ermakova et al. <ref type="bibr" target="#b1">[2]</ref>. The basic ingredients of the track are: Assessments For Task 1, there are qrels for 72 requests (67 requests with at least 1 marginally relevant abstract, 22 requests with at least 5 relevant abstracts).</p><p>We created runs for all the three tasks of the track, which we will discuss in order. Task 1 This task requires ranking scientific abstracts in response to a non-expert, general query prompted by a popular science article. We submitted two runs. The first run, labeled UAms-MF in <ref type="bibr" target="#b1">[2,</ref><ref type="bibr" target="#b2">3]</ref>, is a manual run selecting relevant and accessible results from the top 5 of a vanilla Elastic Search run.</p><p>Our second run, labeled UAms in <ref type="bibr" target="#b1">[2,</ref><ref type="bibr" target="#b2">3]</ref>, is an automatic runs using a reading level/text complexity score as a filter. Specifically, per request and the top 100 result of a vanilla Elastic Search run, we remove 50% of the abstracts with the highest text complexity based on the popular Flesch readability level score.</p><p>Task 2 What concept needs to be explained or rewritten in a given sentence, extracted from a scientific abstract.</p><p>Based on preliminary experiments, our submission also labeled UAms in <ref type="bibr" target="#b1">[2,</ref><ref type="bibr" target="#b3">4]</ref> is using an idf-based term weighting to locate the most rare terms, combined with a simple way to boost particular syntactic categories. Specifically, we used all train and test sentences combined as a reference corpus to calculate document (or rather sentence) frequencies, and use this to rank each term in the source sentence by increasing DF (or decreasing IDF). We include adhoc boost factors for particular part-of-speech, promoting nouns and demoting verbs and adjectives.</p><p>Task 3 Rewrite a sentence from a scientific abstract. This is a post-submission run, hence not evaluated in the track and task overview papers <ref type="bibr" target="#b1">[2,</ref><ref type="bibr" target="#b4">5]</ref>. We use a standard text simplification model, based on the GPT-2 based keep it simple (KiS) model of Laban et al. <ref type="bibr" target="#b5">[6]</ref>. We run a pretrained version of this model available from HuggingFace, 3 in a zero-shot way on both the train and test corpus.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Experimental Results</head><p>In this section, we will present the results of our experiments, in four self-contained subsections following the CLEF 2022 SimpleText Track corpus and tasks.   </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.">Corpus, Context and Requests</head><p>We start with a preliminary analysis of the complexity of the scientific abstracts, in relation to the context and requests. To quantify the complexity, we use the Flesch-Kincaid Grade Level (FKGL) measure based on the lexical and grammatical complexity. This is a simple measure based on word length and sentence length, which may not be the most accurate for a single abstract but a reliable approximation when averaging over larger sets of data. The FKGL score is calibrated to correspond to the readability level suitable for a given school level in the U.S. school system, as shown in Table <ref type="table" target="#tab_1">1</ref>. While literacy levels vary in the population, even among adults, one may assume that an average layperson would have finished compulsory education, corresponding to a high school diploma at a grade level of 12.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.1.">Complexity of the Corpus</head><p>We down-sampled the corpus by taking every 500th article, resulting in an arbitrary sample of 8,513 non-empty abstracts. As shown in Table <ref type="table" target="#tab_2">2</ref>, the average (median) length of the abstracts is 951 (905) tokens, and the average (median) complexity of the abstracts is 14.55 (14.4) FKGL.</p><p>How complex are scientific abstracts? We can immediately confirm that scientific literature is indeed complex: the scale is the U.S. grade levels in years, with 12 being the exit level of compulsory education (high school diploma), hence the observed complexity of 14-15 is translating to students half-way in undergraduate or college education.</p><p>What is the target level of complexity? Recall that the track also provides 40 popular science articles from The Guardian and TechXplore, which are written by professional science journalists for a general audience. As also shown in Table <ref type="table" target="#tab_2">2</ref>, the average (median) length of these articles is 5,504 (5,540) tokens, and the average (median) complexity of the articles is 12.53 (12.7) FKGL, confirming that a FKGL around 12, translating to the readability level of a high school diploma, is appropriate for general citizens.</p><p>Is every single abstract too complex for an average citizen? Figure <ref type="figure" target="#fig_1">1</ref> (left) shows the distribution of FKGL readability levels, which show a striking variation ranging from 5 (elementary school, 10 year old children) to 25 (graduate school domain expert). Figure <ref type="figure" target="#fig_1">1</ref> (right) visualizes this extreme variation, plotted against the length of the abstracts. There is in fact a weak correlation between text complexity and length (r=0.1059, highly significant, regression line with slope 0.0007 in red), but for any length we find abstracts on any level of readability.</p><p>Our analysis confirms the presumption that scientific literature is complex, and a large fraction of abstracts would be very challenging for a layperson. However, our analysis also reveals that a large fraction of abstracts is within the readability levels of most adult citizens.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.2.">Complexity of the Requests</head><p>What subset of abstracts is selected by a general query based on the popular science newspaper articles? We use the default elastic search engine, and retrieve the top 100 scientific articles for each request, and analyze the text complexity of each retrieved abstract. Over the 114 queries, this results in a sample of 11,400 abstracts. As shown also in Table <ref type="table" target="#tab_2">2</ref>, the average (median) length of the retrieved abstracts is 948 (928) tokens, and the average (median) complexity of the abstracts is 13.79 (14.4) FKGL. Hence, the retrieved abstracts are comparable to the corpus statistics, both in terms of length and text complexity, and also the distribution of FKGL (not shown) is very similar.</p><p>Figure <ref type="figure" target="#fig_2">2</ref> shows the distribution of FKGL readability levels over rank of retrieval (left-hand side), and over each individual query (right-hand side). In both cases we see that the standard retrieval engine is completely blind to the text complexity, and exclusively focusing on the topical relevance of the abstract. As a result, for any rank and any topic, we see again a striking variation in FKGL, ranging from 10 (starting high school, 15 year old children) to 20 (doctoral/PhD candidate).</p><p>There are three conclusions based on our analysis: First, a negative result is that we can confirm that scientific abstracts on average are complex, confirming and validating the presumption of laypersons avoiding scientific information. Second, a positive result is that we also found that the variation of complexity is dramatic, and a large fraction of abstracts is within the readability level of an average educated adult citizen. Third, standard search engines are optimized for topic relevance and are completely ignoring secondary aspects such as the reading level of the text.</p><p>Generally, this last finding can immediately explain why laypersons have a disappointing experience when searching academic literature, and explain why they avoid academic sources even when they care about objective evidence. However, note that this finding is also actionable to dramatically improve existing search technology, by explicitly factoring readability into the ranking model, and retrieving all and only relevant abstracts that are not prohibitively complex for an interested outsider.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.">Task 1: Content Selection</head><p>We continue with Task 1, asking to retrieve scientific articles in response to a query based on a popular science article. We submitted two runs, one based on a manual selection of the top 5 results of the default elastic search engine, and an automatic run filtering the top 100 results for the 50% of abstracts with the lowest text complexity based on the readability measure.</p><p>Table <ref type="table" target="#tab_3">3</ref> shows the results of our official submissions for Task 1. All scores are based on the top 5 abstracts retrieved per query, based on a pool of abstracts retrieved by at least 2 submissions, and evaluated on a scale ranging from 0 (irrelevant articles) to 5 (when the abstract and keywords are relevant with the query and the content of the original article). We see that the set of documents in the top 5 of the manual run has a higher level of relevance than the readability level filtered automatic run, and also finds more relevant abstracts per query.</p><p>Using the final qrels, we can also evaluate using familiar search measures, with the resulting evaluation being shown in Table <ref type="table">4</ref>. We also include here the evaluation of the standard Elastic Search for the designated queries (regular queries without quotes). The qrels include 72 topics, but our manual run does not include results for topics in which the top 5 abstracts as returned by Elastic search API were deemed non-relevant for the context of the article. Hence we also evaluate the manual run over the intersection of 52 topics, for which the manual run includes Boolean quantization on various levels of relevance. In the top part of the table, we evaluate on all levels of relevance (with "1" meaning marginally relevant). Here we see very reasonable early precision scores, in particular for the manual run. The high average precision reflects suggests even good performance at higher recall levels. In the middle part of the table, we evaluate on relevance level 2 and higher (with "2" meaning relevant), confirming the superiority of the manual run to return those abstracts with higher levels of relevance. In the bottom part of the table, we evaluate on relevance level 4 and higher (with "4" meaning relevant to the article context), and observe that our automatic run not only returns abstracts that are easier to read, it also outperforms the baseline system in returning those abstracts of direct interest to the article context.</p><p>There are three conclusions based on our analysis. First, for every query and every level of relevance, there exist retrieved documents at a variety of readability levels. Second, a straightforward filter on readability level is able to retrieve abstracts that have a readability suitable for an educated citizen. Third, filtering on readability level leads to a small loss of retrieval effectiveness (as some relevant abstracts have high levels of text complexity), but still obtains a very reasonable performance in particular for retrieving highly relevant articles.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3.">Task 2: Complexity Spotting</head><p>We continue with Task 2, asking to locate the most difficult concepts in a sentence extracted from a potentially relevant abstract, retrieved in response to a general query prompted by a popular science article. We submitted a single run, using an IDF based approach to find the least common term in the sentence, while boosting the most likely part of speech (i.e., nouns or noun phrases).</p><p>Table <ref type="table" target="#tab_4">6</ref> shows the results of our official submission to Task 2, retrieving 263k terms for the 117k sentences in the test corpus. We see that a large fraction of highlighted terms (89%) has correct term limits, which is reassuring as we focused on selecting unigram tokens or words rather than complex concepts or phrases. Term difficulty is judged on a scale ranging from 0 (term needs no explanation) until 7 (impossible to understand). A fair fraction of terms selected (8%) has a high level of difficulty (3 or higher) and the majority of those (5%) a very high level of difficulty (5 or higher), with lower fractions exactly hitting the term limits (5% and 4% respectively).</p><p>At the time of writing, no ground truth is released for Task 2. Fortunately, the organizers released train data in an earlier stage of the track, consisting of 282 sentences with 453 manually extracted complex concepts. By treating every sentence as a "query" and every extracted concept as a "document identifier" (the string of characters after tokenizing and removing white-space), we can calculate standard ranked list measures as shown in Table <ref type="table" target="#tab_5">7</ref>. We observe very reasonable precision scores, for both selecting the longest and for selecting the least common terms. With the IDF approach being more effective than the simple word length. Qualitative inspection reveals that this is particularly caused by abbreviations who tend to be short character strings. The gold standard also selects multi-term concepts, which are systematically missed by our single term or word-based approaches. Still, NDCG (35 to 40%) and MAP (around 30%) remain very reasonable. The NDCG scores reflect the graded concept difficulty level of the ground truth, normalized against the ideal ranking of the most difficult term first, and show impressive performance for these straightforward approaches, and a clearer advantage for the IDF approach over the length based approach. Qualitative inspection reveals that the concepts annotated as the most difficult, have often not the highest lexical complexity, but factor in the importance or centrality of the concept for understanding the sentence and abstract at hand.</p><p>As there are very few terms selected per sentence, leading to a necessarily low precision at 10 or even at 5, we also calculate set based precision, recall, and F1 for the first three results in Table <ref type="table" target="#tab_6">8</ref>. We see here that the set based F1 on the train data is highest after selecting 2 terms per sentence.</p><p>There are three conclusions based on our analysis. First, although spotting complex terms is a hard problem in general, straightforward approaches obtain very reasonable performance in the context of a single sentence. Second, using text statistics such as inverse word frequencies performs better than using only local features such as word length, and are particularly helpful to locate the most difficult terms. Third, lexical complexity as used in readability measures is not enough to locate the ground truth concepts, and we have to combine complexity with importance of the concept in the broader context of the information request. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.4.">Task 3: Text Simplification</head><p>We continue with Task 3, asking to perform text simplification proper, by rewriting a sentence extracted from a potentially relevant abstract, retrieved in response to a general query prompted by a popular science article. We only perform post-submission experiments, based on the zero-shot application of an existing neural text simplification model from <ref type="bibr" target="#b5">[6]</ref>, called the Keep it Simple (KiS) model. The model is based on GPT-medium, using a straightforward unsupervised training task with an explicit loss in terms of fluency, saliency, and simplicity. We are interested in this model as it is fully trained in an unsupervised way, and could be retrained or fine-tuned for the corpus or other academic texts without the need for huge human training data.</p><p>Table <ref type="table" target="#tab_7">9</ref> shows the results of applying the KiS model zero-shot on the train and test data in terms of the generated output. To give an indication of whether the output is indeed simplified, we analyze the Flesch-Kincaid Grade Level (FKGL) of input and output sentences and the resulting compression in token length. We make the following three observations. First, we see a mean and median level of 15 in the scientific abstracts, which we lower by about 3 levels or years of education, with FKGL 12 corresponding to a high school diploma level. Second, we also look at the percentage of sentences where the FKGL is lowered, and see that this is the case in around 80% of sentences. Note that here, the dummy "no change" approach fails miserably, as not a single sentence is simplified. Third, in terms of sentence length, we see no significant compression, as the generated sentences are on par with the input sentences. This may be related to the corpus, as the input sentences from scientific abstracts tend to be not very long with a mean length of 25.8 (train) and 24.2 (test) tokens, and a median length of 24 and 23 respectively. This very significant reduction in text complexity is an encouraging result showing the promise to realize the general aims of the track.</p><p>Table <ref type="table" target="#tab_8">10</ref> shows examples of the generated output for the first three sentences in the train corpus. On the one hand, we see no major issues in language aspects -every generated sentence is grammatical and coherent -and no issues with generating uncontrolled untrue information not contained in the input sentence. On the other hand, we see only very conservative changes, mostly very light editing in terms of deletions and substitutions. While the examples and earlier analysis is showing that we are moving in the right direction, the ground truth is a far more significant simplification. Hence, developing dedicated text simplification approaches for scientific text remains an important open problem.</p><p>For Task 3, we only performed post-submission experiments and our runs were not judged in terms of Lexical complexity, Syntax complexity, and Information loss. Hence we only show  the automatic evaluation based on the human reference simplifications. Table <ref type="table" target="#tab_9">11</ref> shows the automatic evaluation scores for Task 3, using standard SARI and Bleu scores. At the time of writing no test ground truth is available, so we only report scores on the train data. Note we apply a zero-shot model that is neither trained nor fine-tuned in any way on the CLEF SimpleText data, the evaluation on the train corpus is still an independent evaluation of the model's quality. On the train corpus, with a single human simplified reference sentence, the KiS model obtains a Bleu score of 0.2809 and a SARI score of 0.3984. To put this number into perspective, the original paper reports scores in the range of 0.26 to 0.43 on a Wikipedia corpus <ref type="bibr" target="#b6">[7]</ref>. Hence, a 40% SARI score is promising in terms of effectiveness. We also include the dummy "no change" approach making no changes whatsoever, defaulting to returning the input sentence as is. Unlike in machine translation where this would result in very low, if any, token overlap, this proves a competitive approach in terms of the SARI and Bleu scores, as naturally the reference simplification will retain many tokens and n-grams of the original sentence. Recall from above that this no-change approach simplifies not a single sentence, resulting in a 0% of sentences scoring lower on the FKGL scores. This clearly indicates that we need to evaluate multiple aspects to capture the essence of text simplification.</p><p>There are three conclusions based on our analysis. First, an off-the-shelf text simplification model based on GPT-2 is able to rewrite the sentences from academic abstracts with competitive SARI and Bleu scores against high quality human text simplifications. Second, although the model's revisions are conservative, there are no errors introduced and the output is fluent and without loss of information. Third, in terms of readability level, the simplification reduces the level from 15 (college level, undergraduate studies) to 12 (high school diploma, exit level compulsory education), suggesting a readability level suitable for a large fraction of educated citizens.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Discussion and Conclusions</head><p>This paper detailed the University of Amsterdam's participation in the CLEF 2022 SimpleText track.</p><p>We performed an extensive analysis of the corpus in terms of the text complexity, leading to the following findings. First, a negative results is that we can confirm that scientific abstracts on average are complex, validating and confirming the presumption of laypersons avoiding scientific information. Second, a positive result is that we also found that the variation of complexity is dramatic, and a large fraction of abstracts is within the readability level of an average educated adult citizen. Third, standard search engines are optimized for topic relevance and are completely ignoring secondary aspects such as the reading level of the text.</p><p>Next, we made two submissions to the first task retrieving academic abstracts in response to a query based on a popular science article, and found the following. First, for every query and every level of relevance, there exist retrieved documents at a variety of readability levels. Second, a straightforward filter on readability level is able to retrieve abstracts that have a readability suitable for educated citizens. Third, filtering on readability level leads to a small loss of retrieval effectiveness (as some relevant abstracts have high levels of text complexity), but still obtains a very reasonable performance in particular for retrieving highly relevant articles.</p><p>We also participated in the second task, asking to spot difficult terms in sentences from academic abstracts, and made the following observations. First, although spotting complex terms is a hard problem in general, straightforward approaches obtain very reasonable performance in the context of a single sentence. Second, using text statistics such as inverse word frequencies performs better than using only local features such as word length, and are particularly helpful to locate the most difficult terms. Third, lexical complexity as used in readability measures is not enough to locate the ground truth concepts, and we have to combine complexity with importance of the concept in the broader context of the information request.</p><p>Finally, we performed exploratory experiments with a large neural text simplification model based on GPT-2, and arrived at the following conclusions. First, an off-the-shelf text simplification model based on GPT-2 is able to rewrite the sentences from academic abstracts with competitive SARI and Bleu scores against high quality human text simplifications. Second, although the model's revisions are conservative, there are no errors introduced and the output is fluent and without loss of information. Third, in terms of readability level, the simplification reduces the level from 15 (college level, undergraduate studies) to 12 (high school, exit level compulsory education), suggesting a readability level suitable for a large fraction of educated citizens.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: CLEF 2022 SimpleText Corpus: distribution of text complexity in Flesch-Kincaid Grade Levels.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: CLEF 2022 SimpleText Top 100 results: distribution of text complexity in Flesch-Kincaid Grade Levels.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 1</head><label>1</label><figDesc>Text complexity: readability in school grade levels</figDesc><table><row><cell cols="3">Grade Level 1 2 3 4 5 6 7</cell><cell>8</cell><cell cols="2">9 10 11 12 13 14 15 16</cell><cell>17</cell><cell>18 19 20</cell></row><row><cell>School</cell><cell>Elementary</cell><cell cols="2">Jr. High</cell><cell>High School</cell><cell>Undergrad.</cell><cell>Grad.</cell><cell>PhD</cell></row><row><cell></cell><cell>Primary</cell><cell></cell><cell cols="2">Secondary</cell><cell>University</cell><cell></cell><cell>PhD</cell></row><row><cell></cell><cell></cell><cell cols="2">Compulsory</cell><cell></cell><cell cols="2">Higher Edu.</cell></row><row><cell>Age</cell><cell cols="5">6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21</cell><cell>22</cell><cell>23 24 25</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 2</head><label>2</label><figDesc>CLEF 2022 SimpleText Data: Flesch-Kincaid Grade Level.</figDesc><table><row><cell>Data</cell><cell>Sample Size</cell><cell></cell><cell>Length</cell><cell></cell><cell>FKGL</cell></row><row><cell></cell><cell></cell><cell>Mean</cell><cell>Median</cell><cell>Mean</cell><cell>Median</cell></row><row><cell>Corpus (scientific abstracts)</cell><cell>8,513</cell><cell>951</cell><cell>905</cell><cell>14.55</cell><cell>14.40</cell></row><row><cell>News (popular science)</cell><cell>40</cell><cell>5,504</cell><cell>5,540</cell><cell>12.53</cell><cell>12.70</cell></row><row><cell>Retrieved results (top 100)</cell><cell>11,400</cell><cell>948</cell><cell>928</cell><cell>13.79</cell><cell>14.40</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 3</head><label>3</label><figDesc>Results for the SimpleText Task 1: Manual run and filter on median readability score per query.</figDesc><table><row><cell>Run</cell><cell cols="5">Score@5 #Docs@5 Doc Avg #Queries Query Avg</cell></row><row><cell>Manual selection of top 5 Elastic Search</cell><cell>163</cell><cell>54</cell><cell>0.87</cell><cell>99</cell><cell>1.65</cell></row><row><cell>Keep only 50% most readable abstracts</cell><cell>52</cell><cell>17</cell><cell>0.22</cell><cell>40</cell><cell>1.30</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_4"><head>Table 6</head><label>6</label><figDesc>Results for the SimpleText Task 2: Selecting rare terms.</figDesc><table><row><cell>Run</cell><cell>Total</cell><cell cols="2">Evaluated</cell><cell>Score_3</cell><cell></cell><cell>Score_5</cell></row><row><cell></cell><cell></cell><cell></cell><cell>+Limits</cell><cell cols="2">+Limits</cell><cell cols="2">+Limits</cell></row><row><cell cols="2">Selecting rare terms 263,022</cell><cell>1,315</cell><cell>1,175</cell><cell>105</cell><cell>69</cell><cell>60</cell><cell>49</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_5"><head>Table 7</head><label>7</label><figDesc>Evaluation of SimpleText Task 2 (ranked list measures)</figDesc><table><row><cell>Run</cell><cell>MRR</cell><cell></cell><cell>Precision</cell><cell></cell><cell></cell><cell>NDCG</cell><cell></cell><cell>MAP</cell></row><row><cell></cell><cell></cell><cell>5</cell><cell>10</cell><cell>20</cell><cell>5</cell><cell>10</cell><cell>20</cell><cell></cell></row><row><cell>Longest terms</cell><cell>0.3577</cell><cell>0.1128</cell><cell>0.0720</cell><cell>0.0415</cell><cell>0.2994</cell><cell>0.3355</cell><cell>0.3542</cell><cell>0.2863</cell></row><row><cell>Rarest terms</cell><cell>0.3727</cell><cell>0.1340</cell><cell>0.0840</cell><cell>0.0440</cell><cell>0.3462</cell><cell>0.3847</cell><cell>0.3907</cell><cell>0.3088</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_6"><head>Table 8</head><label>8</label><figDesc>Evaluation of SimpleText Task 2 (set measures).</figDesc><table><row><cell>Run</cell><cell></cell><cell>Precision</cell><cell></cell><cell></cell><cell>Recall</cell><cell></cell><cell></cell><cell>F1</cell></row><row><cell></cell><cell>1</cell><cell>2</cell><cell>3</cell><cell>1</cell><cell>2</cell><cell>3</cell><cell>1</cell><cell>2</cell><cell>3</cell></row><row><cell cols="10">Longest terms 0.2376 0.1950 0.1513 0.1690 0.2652 0.3052 0.1903 0.2152 0.1933</cell></row><row><cell>Rarest terms</cell><cell cols="9">0.2305 0.2092 0.1690 0.1727 0.2910 0.3424 0.1883 0.2307 0.2152</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_7"><head>Table 9</head><label>9</label><figDesc>Results for SimpleText Task 3: zero-shot KiS (Keep it Simple) Model<ref type="bibr" target="#b5">[6]</ref>.</figDesc><table><row><cell>Run</cell><cell>Task</cell><cell>Sentences</cell><cell></cell><cell>FKGL</cell><cell></cell><cell cols="2">Compression</cell></row><row><cell></cell><cell></cell><cell></cell><cell>Mean</cell><cell>Median</cell><cell>Ratio</cell><cell>Mean</cell><cell>Median</cell></row><row><cell>No change</cell><cell>Train</cell><cell>648</cell><cell>15.46</cell><cell>15.4</cell><cell>0.00</cell><cell>1.00</cell><cell>1.00</cell></row><row><cell>KiS Model</cell><cell>Train</cell><cell>648</cell><cell>12.78</cell><cell>12.7</cell><cell>0.81</cell><cell>1.15</cell><cell>0.99</cell></row><row><cell>No change</cell><cell>Test</cell><cell>116,763</cell><cell>14.85</cell><cell>14.7</cell><cell>0.00</cell><cell>1.00</cell><cell>1.00</cell></row><row><cell>KiS Model</cell><cell>Test</cell><cell>116,763</cell><cell>12.06</cell><cell>11.9</cell><cell>0.79</cell><cell>1.33</cell><cell>1.01</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_8"><head>Table 10</head><label>10</label><figDesc>Examples simplifications: zero-shot KiS (Keep it Simple) Model<ref type="bibr" target="#b5">[6]</ref> (sentences G11.1_2892036907_{1,2,3})</figDesc><table><row><cell>Input</cell><cell>Output</cell><cell>Reference</cell></row><row><cell>In the modern era of automation</cell><cell>In the modern era of automation</cell><cell>Current academic and industrial re-</cell></row><row><cell>and robotics, autonomous vehicles</cell><cell>and robotics, vehicles are increas-</cell><cell>search is interested in autonomous</cell></row><row><cell>are currently the focus of academic</cell><cell>ingly the focus of academic and in-</cell><cell>vehicles.</cell></row><row><cell>and industrial research</cell><cell>dustrial research.</cell><cell></cell></row><row><cell>With the ever increasing number of</cell><cell>With the ever increasing number of</cell><cell>Drones are increasingly used in</cell></row><row><cell>unmanned aerial vehicles getting</cell><cell>unmanned aerial vehicles getting</cell><cell>the civilian and commercial domain</cell></row><row><cell>involved in activities in the civilian</cell><cell>involved in activities in the civilian</cell><cell>and need to be autonomous.</cell></row><row><cell>and commercial domain, there is</cell><cell>and commercial domain, there is</cell><cell></cell></row><row><cell>an increased need for autonomy in</cell><cell>an increased need for autonomy in</cell><cell></cell></row><row><cell>these systems too</cell><cell>these systems.</cell><cell></cell></row><row><cell>Due to guidelines set by the</cell><cell>Due to guidelines set by the govern-</cell><cell>Governments set guidelines on the</cell></row><row><cell>governments regarding the op-</cell><cell>ments regarding the operation ceil-</cell><cell>operation ceiling of civil drones. So,</cell></row><row><cell>eration ceiling of civil drones,</cell><cell>ing of civil drones, road-tracking</cell><cell>road-tracking based navigation is</cell></row><row><cell>road-tracking based navigation is</cell><cell>based navigation is gaining trac-</cell><cell>attracting interest.</cell></row><row><cell>garnering interest.</cell><cell>tion.</cell><cell></cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_9"><head>Table 11</head><label>11</label><figDesc>Evaluation results for the SimpleText Task 3: zero-shot KiS (Keep it Simple) Model<ref type="bibr" target="#b5">[6]</ref>.</figDesc><table><row><cell>Model</cell><cell>Task</cell><cell>Evaluated</cell><cell>SARI</cell><cell>Bleu</cell><cell></cell><cell cols="2">Precision</cell><cell></cell></row><row><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell>1-gram</cell><cell>2-gram</cell><cell>3-gram</cell><cell>4-gram</cell></row><row><cell>No change</cell><cell>Train</cell><cell>648</cell><cell>0.5571</cell><cell>0.4204</cell><cell>0.6010</cell><cell>0.4531</cell><cell>0.3712</cell><cell>0.3089</cell></row><row><cell>KiS Model</cell><cell>Train</cell><cell>648</cell><cell>0.3984</cell><cell>0.2809</cell><cell>0.4881</cell><cell>0.3176</cell><cell>0.2319</cell><cell>0.1733</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">https://www.theguardian.com/science</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_1">https://techxplore.com/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_2">https://huggingface.co/philippelaban/keep_it_simple</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgments</head><p>This research was conducted as part of the final research projects of the Bachelor in Artificial Intelligence at the University of Amsterdam, We thank the coordinator Dr. Sander van Splunter for his support and flexibility to work around the CLEF deadlines. This research is funded in part by the Netherlands Organization for Scientific Research (NWO CI # CISC.CC.016), and the Innovation Exchange Amsterdam (POC grant). Views expressed in this paper are not necessarily shared or endorsed by those funding the research.</p></div>
			</div>

			<div type="annex">
<div xmlns="http://www.tei-c.org/ns/1.0"><p>We see that the manual run, retrieving between 1 and 5 results per query, has superior early precision, higher than the Elastic Search baseline. We also see that our automatic run obtains very reasonable performance with an NDCG@10 of 38%. The performance in comparison to the original Elastic baseline (scoring an NDCG@10 of 43%) may look unimpressive, as in terms of relevance ranking we do not outperform the baseline. However, recall that our automatic run had a different aim, radically filtering the abstracts to a reading level agreeable with the intended layperson user. Hence we also include the Flesch-Kincaid Grade Level (FKGL) readability scores, and observe that the automatic run is able to return abstracts that are on average 2 years or school levels lower. That is, the reading level of the automatic run is around 12, corresponding to exit level compulsory education, or high school diploma, which would be accessible to the target audience of educated citizens. The baseline approach, in contrast, suggests a reading level requiring college or university education.</p><p>Assuming that users can select from a ranked list, it is of interest to analyze if, and how many, relevant and highly relevant abstracts are in the runs. Table <ref type="table">5</ref> evaluates the runs with</p></div>			</div>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Time&apos;s person of the year: You</title>
		<author>
			<persName><forename type="first">L</forename><surname>Grossman</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">TIME</title>
		<imprint>
			<biblScope unit="volume">168</biblScope>
			<date type="published" when="2006">2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Overview of the CLEF 2022 SimpleText Lab: Automatic simplification of scientific texts</title>
		<author>
			<persName><forename type="first">L</forename><surname>Ermakova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Sanjuan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Kamps</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Huet</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Ovchinnikova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Nurbakova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Araújo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Hannachi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">É</forename><surname>Mathurin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Bellot</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">CLEF&apos;22: Proceedings of the Thirteenth International Conference of the CLEF Association</title>
		<title level="s">Lecture Notes in Computer Science</title>
		<editor>
			<persName><forename type="first">A</forename><surname>Barrón-Cedeño</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><forename type="middle">D S</forename><surname>Martino</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><forename type="middle">D</forename><surname>Esposti</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">F</forename><surname>Sebastiani</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">C</forename><surname>Macdonald</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><surname>Pasi</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Hanbury</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><surname>Potthast</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><surname>Faggioli</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</editor>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<monogr>
		<author>
			<persName><forename type="first">E</forename><surname>Sanjuan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Huet</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Kamps</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Ermakova</surname></persName>
		</author>
		<title level="m">Overview of the CLEF 2022 SimpleText Task 1: Passage selection for a simplified summary</title>
				<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<monogr>
		<author>
			<persName><forename type="first">L</forename><surname>Ermakova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Ovchinnikova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Kamps</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Nurbakova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Araújo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Hannachi</surname></persName>
		</author>
		<title level="m">Overview of the CLEF 2022 SimpleText Task 2: Complexity spotting in scientific abstracts</title>
				<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<monogr>
		<author>
			<persName><forename type="first">L</forename><surname>Ermakova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Ovchinnikova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Kamps</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Nurbakova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Araújo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Hannachi</surname></persName>
		</author>
		<title level="m">Overview of the CLEF 2022 SimpleText Task 3: Query biased simplification of scientific texts</title>
				<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Keep it simple: Unsupervised simplification of multi-paragraph text</title>
		<author>
			<persName><forename type="first">P</forename><surname>Laban</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Schnabel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">N</forename><surname>Bennett</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">A</forename><surname>Hearst</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/2021.acl-long.498</idno>
		<ptr target="https://doi.org/10.18653/v1/2021.acl-long.498" />
	</analytic>
	<monogr>
		<title level="m">ACL/IJCNLP&apos;21: Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, Association for Computational Linguistics</title>
				<imprint>
			<date type="published" when="2021">2021</date>
			<biblScope unit="page" from="6365" to="6378" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Optimizing statistical machine translation for text simplification</title>
		<author>
			<persName><forename type="first">W</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Napoles</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Pavlick</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Callison-Burch</surname></persName>
		</author>
		<idno type="DOI">10.1162/tacl_a_00107</idno>
		<ptr target="https://doi.org/10.1162/tacl_a_00107.doi:10.1162/tacl\_a\_00107" />
	</analytic>
	<monogr>
		<title level="j">Trans. Assoc. Comput. Linguistics</title>
		<imprint>
			<biblScope unit="volume">4</biblScope>
			<biblScope unit="page" from="401" to="415" />
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<ptr target=".org" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Working Notes of CLEF 2022: Conference and Labs of the Evaluation Forum</title>
		<title level="s">CEUR Workshop Proceedings</title>
		<editor>
			<persName><forename type="first">G</forename><surname>Faggioli</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Hanbury</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><surname>Potthast</surname></persName>
		</editor>
		<meeting>the Working Notes of CLEF 2022: Conference and Labs of the Evaluation Forum</meeting>
		<imprint>
			<publisher>CEUR-WS</publisher>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
