=Paper= {{Paper |id=Vol-3180/paper-237 |storemode=property |title=Overview of the CLEF 2022 SimpleText Task 3: Query Biased Simplification of Scientific Texts |pdfUrl=https://ceur-ws.org/Vol-3180/paper-237.pdf |volume=Vol-3180 |authors=Liana Ermakova,Irina Ovchinnikov,Jaap Kamps,Diana Nurbakova,Silvia Araújo,Radia Hannachi |dblpUrl=https://dblp.org/rec/conf/clef/ErmakovaOKNAH22a }} ==Overview of the CLEF 2022 SimpleText Task 3: Query Biased Simplification of Scientific Texts== https://ceur-ws.org/Vol-3180/paper-237.pdf
Overview of the CLEF 2022 SimpleText Task 3: Query
Biased Simplification of Scientific Texts
Liana Ermakova1 , Irina Ovchinnikov2 , Jaap Kamps3 , Diana Nurbakova4 ,
Sílvia Araújo5 and Radia Hannachi6
1
  Université de Bretagne Occidentale, HCTI, France
2
  ManPower Language Solution, Israel
3
  University of Amsterdam, Amsterdam, The Netherlands
4
  University of Lyon, INSA Lyon, CNRS, LIRIS, UMR5205, Villeurbanne, France
5
  Universidade do Minho, CEHUM, 4710-057 Braga, Portugal
6
  Université de Bretagne Sud, HCTI, 56321 Lorient, France


                                         Abstract
                                         This paper presents an overview of the CLEF 2022 SimpleText Task 3 on query biased simplification of
                                         scientific text. After discussing the motivation and general task setup, we detail the exact test collection,
                                         consisting of a train of sentences from scientific abstracts paired with human reference simplified
                                         sentences, and extensive test corpus of sentences with detailed annotations of lexical and syntactic
                                         complexity. We present a detailed analysis of the submitted simplified sentences, and the resulting
                                         evaluation scores.

                                         Keywords
                                         automatic text simplification, science popularization, information distortion, error analysis, lexical
                                         complexity, syntactic complexity




1. Introduction
Digitization and open access have made scientific literature available to every citizen. While this
is an important first step, there are several remaining barriers preventing laypersons to access
the objective scientific knowledge in the literature. In particular, scientific texts are often hard to
understand as they require solid background knowledge and use tricky terminology. Although
there were some recent efforts on text simplification (e.g. [1]), removing such understanding
barriers between scientific texts and general public in an automatic manner is still an open
challenge.
   The CLEF 2022 SimpleText track brings together researchers and practitioners working on
the generation of simplified summaries of scientific texts. It is a new evaluation lab that follows
up the SimpleText-2021 Workshop [2]. The track provides data and benchmarks for discussion
of challenges of automatic text simplification by bringing in the following interconnected tasks:
CLEF 2022: Conference and Labs of the Evaluation Forum, September 5–8, 2022, Bologna, Italy
$ liana.ermakova@univ-brest.fr (L. Ermakova)
€ https://simpletext-project.com/ (L. Ermakova)
 0000-0002-7598-7474 (L. Ermakova); 0000-0003-1726-3360 (I. Ovchinnikov); 0000-0002-6614-0087 (J. Kamps);
0000-0002-6620-7771 (D. Nurbakova); 0000-0003-4321-4511 (S. Araújo)
                                       © 2022 Copyright for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0).
    CEUR
    Workshop
    Proceedings
                  http://ceur-ws.org
                  ISSN 1613-0073
                                       CEUR Workshop Proceedings (CEUR-WS.org)
Table 1
CLEF 2022 SimpleText official run submission statistic
 Team                           Task 1             Task 2              Task 3             Total runs
 aaac                                              1     (1 updated)                              1
 CLARA-HD [6]                                                           1                         1
 CYUT Team2 [7]                 1                                       1                         2
 HULAT-UC3M [8]                                                        10   (4 updated)          10
 LEA_T5 [9]                                        1                    1                         2
 NLP@IISERB [10]                3   (3 updated)                                                   3
 PortLinguE [11]                                                        1   (1 updated)           1
 SimpleScientificText [12]                         1     (1 updated)                              1
 UAms [13]                      2                  1                                              3
 Total runs                     6                  4                   14                        24



Task 1: What is in (or out)? Select passages to include in a simplified summary, given a
     query.

Task 2: What is unclear? Given a passage and a query, rank terms/concepts that are required
     to be explained for understanding this passage (definitions, context, applications,..).

Task 3: Rewrite this! Given a query, simplify passages from scientific abstracts.

This paper focuses on the third task of text simplification proper. We refer for details of the
other tasks to the overview papers of Task 1 [3] and Task 2 [4], or the Track overview paper [5].
   In the CLEF 2022 edition of SimpleText, a total of 62 teams registered for the SimpleText track.
A total of 40 users downloaded data from the server. A total of 9 distinct teams submitted 24
runs, of which 10 runs were updated. The details of statistics on runs submitted for shared tasks
are presented in Table 1. As for the third task, a total of 14 runs from five teams were submitted.
   This introduction is followed by Section 2 presenting the text simplification task with the
datasets and evaluation metrics used. In Section 3, we discuss the results of the official submis-
sions. We end with Section 4 discussing the results and findings, and lessons for the future.


2. CLEF 2022 SimpleText Task 3 Test Collection
In this section, we discuss the third task about text simplification proper, rewriting an extracted
sentence from a scientific abstract, addressing the task:

      Given a query, simplify passages from scientific abstracts.

   The goal of this task is to provide a simplified version of text passages (sentences) with regard
to a query. Participants were provided with queries and abstracts of scientific papers. The
abstracts could be split into sentences. The simplified passages were evaluated manually in
terms of the produced errors as follows.
Table 2
SimpleText Task 3: Statistics of the number of evaluated sentences per query
 Query                          # Distinct source sentences       # Distinct simplified sentences
  1   digital assistant                                   370                               1,280
  2   conspiracy theories                                 195                                 398
  3   end to end encryption                                55                                 102
  4   imbalanced data                                      55                                  87
  5   genetic algorithm                                    51                                  85
  6   quantum computing                                    51                                  85
  7   qbit                                                 50                                  76
  8   quantum applications                                 42                                  73
  9   cyber-security                                       28                                  47
 10   fairness                                             18                                  22
 11   crowsourcing                                         14                                  21


2.1. Train Data
As for Task 2: What is unclear?, we provided a parallel corpus of simplified sentences from two
domains: Medicine and Computer Science. As previously, we use scientific abstracts from the
DBLP Citation Network Dataset for Computer Science and Google Scholar and PubMed articles
on muscle hypertrophy and health Medicine [14, 15].
   Text passages issued from abstracts on computer science were simplified by either a master
student in Technical Writing and Translation or a pair of experts: (1) a computer scientist and
(2) a professional translator, English native speaker but not specialist in computer science [15].
Each passage was discussed and rewritten multiple times until it became clear for non-computer
scientists. Medicine articles were annotated by a master student in Technical Writing and
Translation specializing in this domain. Sentences were shortened, excluding every detail that
was irrelevant or unnecessary to the comprehension of the study, and rephrased, using simpler
vocabulary. If necessary, concepts were explained.
   We provided 648 parallel sentences in total.

2.2. Test Data
We used the same 116,763 sentences retrieved by the ElasticSearch engine from the DBLP
dataset according to the queries as for Task 2. We manually evaluated 2,276 pairs of sentences
for 11 queries. For the query Digital assistant we took the first 1,000 sentences retrieved by
ElasticSearch. We pool source sentences coupled with their simplified versions submitted by
all participants for all these queries. We ensured that for each evaluated source sentence the
pool contained the results of all participants. The detailed statistics of the number of evaluated
sentences per query for Task 3 are given in Table 2.
2.3. Input and Output Format
The input train and and the test data were provided in JSON and CSV formats with the following
fields:

snt_id a unique passage (sentence) identifier.

source_snt passage text.

doc_id a unique source document identifier.

query_id a query ID.

query_text simplification should be done with regard to this query.

Input example (JSON format):
{"snt_id":"G11.1_2892036907_2", "source_snt":"With the ever increasing number of
˓→  unmanned aerial vehicles getting involved in activities in the civilian and
˓→  commercial domain, there is an increased need for autonomy in these systems
˓→  too.", "doc_id":2892036907, "query_id":"G11.1", "query_text":"drones"}

  Participants were asked to provide a list of terms to be contextualized in a JSON format or a
tabulated file TSV (for manual runs) with the following fields:

run_id Run ID starting with (team_id)_(task_3)_(name).

manual Whether the run is manual {0, 1}.

snt_id a unique passage (sentence) identifier from the input file.

simplified_snt Text of the simplified passage .

Output example (JSON format):
{"run_id":"BTU_task_3_run1", "manual":1, "snt_id":"G11.1_2892036907_2",
˓→  "simplified_snt":"Drones are increasingly used in the civilian and commercial
˓→  domain and need to be autonomous."}


2.4. Evaluation metrics
We filtered out the simplified sentences identical to the source ones and the truncated sim-
plified sentences by keeping only passages matching the regular expression (valid snippets):
.+[?.!"\']\s*\$'
  Professional linguists manually annotated simplifications provided with regard to a query
according to the following criteria. We evaluated binary errors:

    • Incorrect syntax;
    • Unresolved anaphora due to simplification;
    • Unnecessary repetition/iteration (lexical overlap);
    • Spelling, typographic or punctuation errors.
The lexical and syntax complexity of the produced simplifications were assessed on an absolute
scale, value 1 referring to a simple output sentence regardless of the complexity of the source
one, 7 corresponding to a complex one. Lexical complexity is mostly identical to that presented
in Task 2, see the Track overview [5] and Task overview [4].
   We consider syntax complexity based on syntactic dependencies, their length and depth.
The dependency trees reveal latent complications for reading and understanding text; thus,
psycho-linguists consider the syntactic dependencies to be a relevant tool to evaluate text
readability [16]. The depth and length of the syntactic chains we interpret according to [16].
   We evaluate syntax complexity as follows:
   1. Simple sentence (without negation / passive voice): Over Facebook, we find many interac-
      tions.
   2. Simple sentence with negation / passive voice (e.g. Many interactions were found over Face-
      book) or Simple sentences with syntactic constructions that show chains of dependency
      and shallow embedding depth (e.g. Over Facebook, we find many interactions between
      public pages and both political wings.)
   3. Simple sentences with long chains of dependency and shallow embedding depth, with
      syntactic constructions like complex object, gerund construction, etc. (e.g. Despite the
      enthusiastic rhetoric about the so-called collective intelligence, conspiracy theories have
      emerged.) or Short complex or compound sentence (e.g. We propose a novel approach that
      was used in terms of information theory.)
   4. Simple sentences with long chains of dependency and deep embedding depth, with
      syntactic constructions like complex object, gerund construction, etc. (e.g. Over Facebook,
      we find many interactions between public pages for military and veterans, and both sides of
      the political spectrum) or Complex or compound sentence that contains long chains of
      dependency and deep embedding depth;
   5. Simple sentences with long chains of dependency and deep embedding depth, with several
      syntactic constructions like complex object, gerund construction, etc. or & Complex or
      compound sentence that contains long chains of dependency and deep embedding depth;
   6. Complex or compound sentences that contain long chains of dependency and deep
      embedding depth along with complex object, gerund construction, etc. or Simple sentence
      that contains modifications, topicalization, parenthetical constructions: Moreover, we
      measure the effect of 4709 evidently false information (satirical version of conspiracist stories)
      and 4502 debunking memes (information aiming at contrasting unsubstantiated rumors) on
      polarized users of conspiracy claims.
   7. Long complex or compound sentences that contain several clauses of different types, long
      chains of dependency and deep embedding depth along with complex object, gerund
      construction, etc.
  We evaluate the information quality of the simplified snippet based on its content and
readability. Transformation of information from the source snippet brings in omission of details,
insertion of basic terms to explain particular terminology and complex concepts, reference to
resources. Due to necessary insertions and references, the simplified snippets often contain
more words and syntactic constructions as compared to their source. Nevertheless, the goal is to
reduce lexical and syntax complexity in the extended simplified snippets. In case the simplified
snippet lacks information mentioned in the source, we evaluate the degree of the information
loss. Irrelevant insertions, iterations and wordy statements in the extended simplified snippet we
consider as a misrepresentation or distortion of source information when a reader experiences
difficulties in processing source content due to wordiness of the loosely structured simplified
snippet.
   We assessed the information loss severity during the simplification with regard to a given
query on the scale from 1 to 7, where 1 corresponds to unimportant information loss while 7
refers to the most severe information distortion. We consider the information loss as a kind of
information damage even if the information in the simplified text contains the information of
the source passage but has some insertions, which impedes perception of the content.
   We distinguish the following 11 types of misrepresentation of source information. Our classi-
fication leans on the error typology in machine translation [17, 18]. The simplified snippet often
combines several types of distortion, e.g. omission and ambiguity. Nevertheless, we observed
many instances of small distortions that severely diminish the quality of the simplification;
therefore, we need to explain each type providing the clear and transparent instances. Our eval-
uation of the value of the information distortion leans on the calculation of the information loss
and assessment of the diminished readability of the simplified snippet that generates difficulties
in text semantic processing by readers.
   We distinguish the following types of information distortion:
   1. Style (distortion severity 1)
      Source snippet: In order to facilitate knowledge transfer between specialists and generalists
      and between experts and novices, and to promote interdisciplinary communication, there is a
      need to provide methods and tools for doing so.
      Simplified snippet: There is a need to provide methods and tools for doing so. In order to
      facilitate knowledge transfer between specialists and generalists and between experts and
      novices, we need to promote interdisciplinary communication. We need to make it easier for
      people to share their knowledge with each other.
      Comment: Deviations from the style norms do not lead to information loss; however,
      they diminish the quality of text structure and affect readers’ assessment of the text and
      its content.
   2. Insertion of unnecessary details with regard to a query(distortion severity 1)
      Source snippet: In the clinical setting, availability of needed information can be crucial
      during the decision-making process.
      Simplified snippet: availability of needed information can be crucial during the decision-
      making process. In the clinical setting, needed information is often difficult to come by. For
      confidential support call the Samaritans on 08457 90 90 90 or visit a local Samaritans branch,
      see www.samaritans.org for details
      Comment: The simplified snippet often contains more information than the source since
      the terminology is needed to be explained. An irrelevant insertion does not lead to the
      loss of information; however, it may bring in diminishing of the text readability and
      generate discomfort during text perception. The irrelevant reference to the support in the
      simplified snippet does not clarify the source. The source does not need any simplification.
3. Redundancy (without lexical overlap) (distortion severity 2)
   Source snippet: The capability to get updated information and news is an important and
   decisive factor in business and finance.
   Simplified snippet: The capability to get updated information and news is an important and
   decisive factor in business and finance. The ability to get updates on the latest news is also
   an important factor in the success of a business or finance company. For more information,
   visit CNN.com/News.
   Comment: Irrelevant insertions, iterations and wordy statements in the extended simpli-
   fied snippet we consider as a misrepresentation or distortion of source information when
   a reader may misunderstand source content due to wordiness of the loosely structured
   simplified snippet.
4. Insertion of false or unsupported information (distortion severity 3)
   Source snippet: The proposed method leads to not only faster running time but also efficient
   text localization.
   Simplified snippet: The proposed method leads to not only faster running time but also
   efficient text localization. The proposed method is based on the idea that text should be
   written in a single block of text, rather than a series of blocks of text. The method is being
   developed by the University of California, San Diego
   Comment: False and unsupported information is inserted in the simplified snippet because
   of external links of the source snippet to the open Web resources. False information
   confuses the readership, thus it is inappropriate in simplified texts.
5. Omission of essential details with regard to a query (distortion severity 4)
   Source snippet: In short, this thesis aims to repatriate young people’s web practices from
   the sterile, positivist methods space of questionnaires and tests of digital literacy to social
   contexts of everyday life.
   Simplified snippet: In short, this thesis aims to repatriate young people’s web practices from
   the sterile, and tests of digital literacy.
   Comment: Simplified texts regularly omit details and minor concepts since the objective
   of the simplification is to facilitate perception and processing of the main idea and novel
   information of the source. Therefore, distinguishing the essential details and concepts
   from the minor ones plays a crucial role in the text simplification. Omission of the essential
   details leads to the lack of information and novelty of the message. The simplified snippet
   lacks the essential information about the direction where the young people’s web practice
   needs to be repatriated (to social contexts of everyday life).
6. Overgeneralization (distortion severity 5)
   Source snippet: Online Social Networks explode with activity whenever a crisis event takes
   place.
   Simplified snippet: It explodes when a crisis event takes place.
   Comment: In the simplified snippet the subject of the source snippet Online Social Networks
   is omitted being substituted by the pronoun It; also the essential detail with activity
   is omitted that brings in overgeneralization: the statement refers to all cases of the
   dysfunctions (server, browser, users’ activity, moderators’ restrictions, etc.)
7. Oversimplification (distortion severity 5)
    Source snippet: If we accept the current orthodoxy and then blame the Web we offer a tech-
    nological determinist explanation of reality: technology produces misinformed populations.
    Simplified snippet: If we accept the current orthodoxy and then blame the Web we have an
    explanation of reality.
    Comment: Oversimplification appears in the shortened simplified snippets when source
    utterance is transformed into a trivial statement or even a claim. The simplified snippet
    claims that we have an explanation of reality thanks to acceptance of the current orthodoxy
    and accusation of the Web. Meanwhile, the source snippet discusses the technological
    determinist explanation of reality. The omission of the essential details leads to the
    oversimplified statement that cannot explain the reality of the technological epoch.
 8. Topic shift (distortion severity 5)
    Source snippet: global warming induced by chemtrails or the link between vaccines and
    autism – find on the Web a natural medium for their dissemination.
    Simplified snippet: The link between vaccines and autism – can be found on the Web a
    natural medium for changing.
    Comment: Topic shift is revealed in a substitution of the source topic by omitting its part
    or selecting a wrong basic word to replace the peculiar term in the source. The source
    snippet lost the essential part of its topic (global warming induced by chemtrails) during
    the simplification process; moreover, the simplification resulted in the inappropriate
    syntactic structure of the snippet.
 9. Contra sense / contradiction (distortion severity 6)
    Source snippet: In this paper we discuss architectural design issues and trade-offs in con-
    nection with our experiences porting our agent-based platform, Opal, to the Sharp Zaurus
    personal digital assistant (PDA).
    Simplified snippet: The Sharp Zaurus is a personal digital assistant (PDA) developed by
    Sharp. It is based on the Opal agent-based platform. We discuss architectural design issues
    and trade-offs in connection with our experiences porting Opal to the Zaurus PDA.
    Comment: Contradictions in simplified snippets appear due to elimination of essential
    concepts or interrelations among concepts, omission of significant details, and transfor-
    mation of the semantic structure of the source snippet. The simplified snippet mentions
    agent-based platform Opal as the basis for the Sharp Zaurus, but at the same time claims
    that Opal was ported to the Sharp Zaurus. The source snippet But the new phenomena, the
    non-agenda ownership, overcome any ideological influence, especially under the conditions
    of punishment mechanism applied to old politicians lost its semantic structure since the
    concepts ideological influence and punishment mechanism were eliminated in the process
    of its simplification. Thus, the simplified snippet But the new phenomena, the ownership
    of the non-agenda, had a lot of influence on old politicians lacks any explanation how the
    non-agenda ownership is related to old politicians and why they are influence by the new
    phenomena.
10. Ambiguity (distortion severity 6)
    Source snippet: The experimental results show that 3D maps with texture on mobile phone
    display size, and 3D maps without texture on PDA display size are superior to 2D maps in
    search time and error rate.
      Simplified snippet: 3D maps with texture on mobile phone display size are superior to 2D
      maps in search time and error rate. The experimental results show that 3D maps without
      texture on PDA display size were superior to those with texture. The results were published
      in the journal 3D Maps.
      Comment: Ambiguity presupposes that a statement has several equiprobable interpre-
      tations. The instance of the ambiguous simplified snippet in Table X lacks a key to
      understand whether the 3D maps without texture outperform those with texture or not.
      Ambiguity often appears due to syntactic simplification of the source. In the source, the
      clause changes in the strength of competition also reveal key asymmetrical differences is re-
      placed by shorter clause but they do not have any biases that produces ambiguity: whether
      evidence corresponds to reality or not. The source clarifies the differences between two
      political parties: Though both Republicans and Democrats show evidence of implicit biases,
      changes in the strength of competition also reveal key asymmetrical differences however, the
      simplified snippet doubts the reliability of the evidence: Both Republicans and Democrats
      show evidence of biases, but they do not have any biases. Readers of the simplified snippet
      are unable to resolve the ambiguity.
  11. Nonsense (distortion severity 7)
      Source snippet: The large availability of user provided contents on online social media
      facilitates people aggregation around shared beliefs, interests, worldviews and narratives
      Simplified snippet: The large amount of user provided contents on online social media is
      called aggregation
      Comment: The source snippet was transformed into a simple sentence. The transforma-
      tion brings in erroneous usage of the word aggregation that leads to the loss of meaning
      of the whole sentence. Instead of the original statement about accessibility of the social
      or public media on the Web, which facilitates dissemination of fake news and rumors,
      the simplified snippet claims that there is an opportunity to find a resource to read about
      fake news and rumors.
  The final ranking for Task 3 was done by the average harmonic mean of normalized opposite
values of Lexical Complexity (LC), Syntactic Complexity (SC) and Distortion Level (DL) as follows:
                                                        3
                                   𝑠𝑖 =     7      7      7                                    (1)
                                          7−LC + 7−SC + 7−DL
                                               {︃
                                           ∑︀    𝑠𝑖 ,       if No Error
                                             𝑖
                                                 0,         otherwise
                                 Score =                                                    (2)
                                                    𝑛
In Equation 2, variable 𝑛 refers to the total number of judged snippets and No Error means
that the snippet 𝑖 does not have any of Uncorrect syntax, Unresolved anaphora, nor Unnecessary
repetition/iteration error.


3. SimpleText Task 3 Results
In this section we discuss the results for the official submissions to the Task 3.
Table 3
SimpleText Task 3: General results of official runs




                                                                                                                         Unresolved Anaphora




                                                                                                                                                                            Lexical Complexity
                                                                                                                                                        Syntax Complexity
                                                                                                      Uncorrect Syntax




                                                                                                                                                                                                 Information Loss
                                                                          Length Ratio
                                 Unchanged



                                             Truncated




                                                                                          Evaluated




                                                                                                                                               Minors
                                                                 Longer
                                                         Valid
                         Total
Run




CLARA-HD        116,763    128   2,292 111,627 201 0.61                                  851 28                                   3 68 2.10 2.42 3.84
CYUT Team2      116,763    549 101,104 111,818    49 0.81                                126 1                                       32 2.25 2.30 2.26
PortLinguE_full 116,763 42,189     852 111,589 3,217 0.92                                564 7                                        5 2.94 3.06 1.50
PortLinguE_run1   1,000    359       7     970    30 0.93                                 80 1                                           3.63 3.57 2.27
lea_task3_t5     23,360     52 23,201 22,062      24 0.35                                   .  .                                   .   .     .    .    .
HULAT-UC3M01      1,000       .     13     973 968 2.46                                   95 10                                   1 20 4.69 3.69 2.20
HULAT-UC3M02      2,001      3      58   1,960 1,920 2.53                                205 10                                   1 37 3.60 3.53 2.34
HULAT-UC3M03      1,000      2      13     958 966 2.53                                     .  .                                   .   .     .    .    .
HULAT-UC3M04      2,000       .     33   1,827 1,957 37                                     .  .                                   .   .     .    .    .
HULAT-UC3M05      2,000       .     56   1,921 1,918 2.38                                   .  .                                   .   .     .    .    .
HULAT-UC3M06      2,000       .     47   1,976 1,921 2.45                                   .  .                                   .   .     .    .    .
HULAT-UC3M07      1,000       .     56     970 972 2.43                                     .  .                                   .   .     .    .    .
HULAT-UC3M08      2,000       .     62   1,964 1,919 2.59                                   .  .                                   .   .     .    .    .
HULAT-UC3M09      2,000       .    170   1,964 1,904 2.15                                   .  .                                   .   .     .    .    .
HULAT-UC3M10      2,000       .    215   1,963 1,910 2.13                                   .  .                                   .   .     .    .    .


   A total of 5 different teams submitted 14 runs (5 runs were updated). Absolute number of
errors and average Lexical Complexity, Syntax Complexity and Information Loss are provided in
Tables 3 and 4. The final ranking for Task 3 is given in Table 5. We removed all runs with the 0
score.
   Very interesting partial runs were provided by the HULAT-UC3M team as the generated
simplifications provided the explanations of difficult terms. However, HULAT-UC3M’s 8 runs
over 10 were not in the pool with selected topics. Thus, we provided only automatic evaluation
results. The HULAT-UC3M’s runs provide clear evidence of the interconnection of tasks 2 and 3.


4. Conclusion
This paper presented the overview of the CLEF 2022 SimpleText Task 3 on simplifying sentences
in scientific abstracts, retrieved in response to a queries based on popular science articles.
   We created a corpus of sentences extracted from the abstracts of scientific publications. In
contrast to previous work, we evaluate simplification in terms of lexical and syntax complexity
combining with error analysis. We introduced a new classification of information distortion
types for automatic simplification and we annotated the collected simplifications according to
this error classification. Recent pandemics have shown that simplification can be modulated by
Table 4
SimpleText Task 3: Information distortion in evaluated runs




                                                                                                             Omission Of Essential Details




                                                                                                                                                                                        Unsupported Information

                                                                                                                                                                                                                  Unnecessary Details
                                                                                                                                             Overgeneralization


                                                                                                                                                                   Oversimplification
                                                                              Wrong Synonym




                                                                                                                                                                                                                                        Redundancy
                                                 Contresens

                                                                Topic Shift




                                                                                              Ambiguity
                                     Non-Sense
                        Evaluated




                                                                                                                                                                                                                                                     Style
Run
CLARA-HD              851           162          68             37            20              80            314                              59                   203                   26                        10                    29           13
CYUT Team2            126             2           1               .             .              4             42                               4                     5                     .                         .                     .           4
PortLinguE_full       564             9           3              4             3              19             94                               9                    13                    2                         2                     5            1
PortLinguE_run1        80              .           .             1              .               .            27                               5                     2                     .                         .                     .            .
lea_task3_t5             .             .           .              .             .               .              .                               .                     .                    .                         .                     .            .
HULAT-UC3M01           95             1           7              2              .              5              2                                .                    1                    5                        38                    36             .
HULAT-UC3M02          205             4           9              4              .              9              4                                .                     .                  12                        72                    61            1


Table 5
SimpleText Task 3: Ranking of official submissions on combined score
                                                              Run                                         Score
                                                              PortLinguE_full                             0.149
                                                              CYUT Team2                                  0.122
                                                              CLARA-HD                                    0.119


political needs and the scientific information can be distorted. Thus, in contrast to previous
work, we evaluated the simplifications in terms of information distortion.
   For next year, we plan to continue the Task 3 setup, continuing the detailed manual annota-
tions of samples, but also working on automatic metrics that best reflect the insights of this
year’s analysis. This year, the HULAT-UC3M team submitted runs which combine tasks 2 and 3
which demonstrates strong interconnection of the tasks as often the terminology cannot be
removed nor simplified but it needs to be explained to a reader.

Acknowledgments
We like to acknowledge the support of the Lab Chairs of CLEF 2022, Allan Hanbury and Martin Potthast, for
their help and patience. Special thanks to the University Translation Office of the Université de Bretagne
Occidentale, and to Nicolas Poinsu and Ludivine Grégoire for their major impact in the train data construction
and Léa Talec-Bernard and Julien Boccou for their help in evaluation of participants’ runs. We thank Josiane
Mothe for reviewing papers. We also thank Alain Kerhervé, and the MaDICS (https:// www.madics.fr/ ateliers/
simpletext/ research group.
References
 [1] M. Maddela, F. Alva-Manchego, W. Xu, Controllable Text Simplification with Explicit Paraphrasing
     (2021). URL: http://arxiv.org/abs/2010.11004.
 [2] L. Ermakova, P. Bellot, P. Braslavski, J. Kamps, J. Mothe, D. Nurbakova, I. Ovchinnikova, E. Sanjuan,
     Text Simplification for Scientific Information Access: CLEF 2021 SimpleText Workshop, in: Advances
     in Information Retrieval - 43nd European Conference on IR Research, ECIR 2021, Lucca, Italy, March
     28 – April 1, 2021, Proc., Lucca, Italy, 2021.
 [3] E. SanJuan, S. Huet, J. Kamps, L. Ermakova, Overview of the CLEF 2022 SimpleText Task 1: Passage
     selection for a simplified summary, in: [19], 2022.
 [4] L. Ermakova, I. Ovchinnikova, J. Kamps, D. Nurbakova, S. Araújo, R. Hannachi, Overview of the
     CLEF 2022 SimpleText Task 2: Complexity spotting in scientific abstracts, in: [19], 2022.
 [5] L. Ermakova, E. SanJuan, J. Kamps, S. Huet, I. Ovchinnikova, D. Nurbakova, S. Araújo, R. Hannachi,
     É. Mathurin, P. Bellot, Overview of the CLEF 2022 SimpleText Lab: Automatic simplification of
     scientific texts, in: A. Barrón-Cedeño, G. D. S. Martino, M. D. Esposti, F. Sebastiani, C. Macdonald,
     G. Pasi, A. Hanbury, M. Potthast, G. Faggioli, N. Ferro (Eds.), CLEF’22: Proceedings of the Thirteenth
     International Conference of the CLEF Association, Lecture Notes in Computer Science, Springer,
     2022.
 [6] A. Menta, A. Garcia-Serrano, Controllable Sentence Simplification Using Transfer Learning, in:
     Proceedings of the Working Notes of CLEF 2022 - Conference and Labs of the Evaluation Forum,
     Bologna, Italy, September 5th - to - 8th, 2022, CEUR Workshop Proceedings, CEUR-WS.org, Bologna,
     Italy, 2022.
 [7] S.-H. Wu, H.-Y. Huang, CYUT Team2 SimpleText Shared Task Report in CLEF-2022, in: Proceedings
     of the Working Notes of CLEF 2022 - Conference and Labs of the Evaluation Forum, Bologna, Italy,
     September 5th - to - 8th, 2022, CEUR Workshop Proceedings, CEUR-WS.org, Bologna, Italy, 2022.
 [8] A. Rubio, P. Martínez, HULAT-UC3M at SimpleText@CLEF-2022: Scientific text simplification
     using BART, in: Proceedings of the Working Notes of CLEF 2022 - Conference and Labs of the
     Evaluation Forum, Bologna, Italy, September 5th - to - 8th, 2022, CEUR Workshop Proceedings,
     CEUR-WS.org, Bologna, Italy, 2022.
 [9] T.-B. Talec-Bernard, Is Using an AI to Simplify a Scientific Text Really Worth It?, in: Proceedings of
     the Working Notes of CLEF 2022 - Conference and Labs of the Evaluation Forum, Bologna, Italy,
     September 5th - to - 8th, 2022, CEUR Workshop Proceedings, CEUR-WS.org, Bologna, Italy, 2022.
[10] S. Saha, D. Roy, B. Y. Goud, C. S. Reddy, T. Basu, NLP-IISERB@Simpletext2022: To Explore the
     Performance of BM25 and Transformer Based Frameworks for Automatic Simplification of Scientific
     Texts, in: Proceedings of the Working Notes of CLEF 2022 - Conference and Labs of the Evaluation
     Forum, Bologna, Italy, September 5th - to - 8th, 2022, CEUR Workshop Proceedings, CEUR-WS.org,
     Bologna, Italy, 2022.
[11] J. Monteiro, M. Aguiar, S. Araújo, Using a Pre-trained SimpleT5 Model for Text Simplification in a
     limited Corpus, in: Proceedings of the Working Notes of CLEF 2022 - Conference and Labs of the
     Evaluation Forum, Bologna, Italy, September 5th - to - 8th, 2022, CEUR Workshop Proceedings,
     CEUR-WS.org, Bologna, Italy, 2022.
[12] H. Jianfei, M. Jin, Assembly Models for SimpleText Task 2: Results from Wuhan University Research
     Group, in: Proceedings of the Working Notes of CLEF 2022 - Conference and Labs of the Evaluation
     Forum, Bologna, Italy, September 5th - to - 8th, 2022, CEUR Workshop Proceedings, CEUR-WS.org,
     Bologna, Italy, 2022.
[13] F. Mostert, A. Sampatsing, M. Spronk, J. Kamps, University of Amsterdam at the CLEF 2022
     SimpleText Track, in: Proceedings of the Working Notes of CLEF 2022 - Conference and Labs of
     the Evaluation Forum, Bologna, Italy, September 5th - to - 8th, 2022, CEUR Workshop Proceedings,
     CEUR-WS.org, Bologna, Italy, 2022.
[14] L. Ermakova, P. Bellot, P. Braslavski, J. Kamps, J. Mothe, D. Nurbakova, I. Ovchinnikova, E. SanJuan,
     Overview of SimpleText 2021 - CLEF Workshop on Text Simplification for Scientific Information
     Access, in: K. S. Candan, B. Ionescu, L. Goeuriot, B. Larsen, H. Müller, A. Joly, M. Maistro, F. Piroi,
     G. Faggioli, N. Ferro (Eds.), Experimental IR Meets Multilinguality, Multimodality, and Interaction,
     Lecture Notes in Computer Science, Springer International Publishing, Cham, 2021, pp. 432–449.
[15] L. Ermakova, P. Bellot, J. Kamps, D. Nurbakova, I. Ovchinnikova, E. SanJuan, E. Mathurin, S. Araújo,
     R. Hannachi, S. Huet, N. Poinsu, Automatic Simplification of Scientific Texts: SimpleText Lab at
     CLEF-2022, in: M. Hagen, S. Verberne, C. Macdonald, C. Seifert, K. Balog, K. Nørvåg, V. Setty (Eds.),
     Advances in Information Retrieval, volume 13186, Springer International Publishing, Cham, 2022,
     pp. 364–373.
[16] R. Futrell, E. Gibson, H. J. Tily, I. Blank, A. Vishnevetsky, S. T. Piantadosi, E. Fedorenko, The Natural
     Stories corpus: a reading-time corpus of English texts containing rare syntactic constructions, Lan-
     guage Resources and Evaluation 55 (2021) 63–77. URL: https://doi.org/10.1007/s10579-020-09503-7.
[17] I. Ovchinnikova, Impact of new technologies on the types of translation errors, in: CEUR Workshop
     Proceedings, 2020.
[18] A. Lommel, A. Görög, A. Melby, H. Uszkoreit, A. Burchardt, M. Popović, Harmonised Metric, Quality
     Translation 21 (QT21) (2015). URL: https://www.qt21.eu/wp-content/uploads/2015/11/QT21-D3-1.
     pdf.
[19] G. Faggioli, N. Ferro, A. Hanbury, M. Potthast (Eds.), Proc. of the Working Notes of CLEF 2022:
     Conference and Labs of the Evaluation Forum, CEUR Workshop Proceedings, 2022.