<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>LLM-Guided Planning and Summary-Based Scientific Text Simplification: DS@GT at CLEF 2025 SimpleText</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Krishna Chaitanya Marturi</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Heba H. Elwazzan</string-name>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Georgia Institute of Technology</institution>
          ,
          <addr-line>North Ave NW, Atlanta, GA 30332</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>In this paper, we present our approach for the CLEF 2025 SimpleText Task 1, which addresses both sentence-level and document-level scientific text simplification. For sentence-level simplification, our methodology employs large language models (LLMs) to first generate a structured plan, followed by plan-driven simplification of individual sentences. At the document level, we leverage LLMs to produce concise summaries and subsequently guide the simplification process using these summaries. This two-stage, LLM-based framework enables more coherent and contextually faithful simplifications of scientific text.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;LLMs</kwd>
        <kwd>Text Simplification</kwd>
        <kwd>CLEF 2025</kwd>
        <kwd>CEUR-WS</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>In the past decade or so, a new form of learning has emerged. Instead of science being only accessible
through journals or formal education, the general public is now privy to an enormous wealth of material
through the Internet. Whether it be directed self-studying, or casual social media perusal, nearly
everyone is now able to access scientific information. An important caveat remains, however, and
that is when a resource is accessible for free and without rigorous moderation, misinformation will
invariably run rampant. This is why now more than ever, the need for reliable, easy-to-understand
scientific-based text and content has taken the forefront.</p>
      <p>
        This is where automated text simplification comes in. With the volume of scientific text at hand,
rewriting the same exact content in layman terms manually is intractable. Resources have been
expended towards automating this task, and over the course of 20 years, automatic text simplification
has progressed significantly [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], reaching a critical point with the development of Natural Language
processing techniques, and more recently, with the widespread use of LLMs.
      </p>
      <p>The capabilities of Large Language Models have made them a game-changer for automatic text
simplification. Unlike previous methods, LLMs can achieve a deeper semantic interpretation of source
text, allowing them to not only simplify vocabulary and syntax but also to summarize and restructure
information for clarity. Though their internal representations of knowledge are opaque, their practical
application as a powerful tool for generating simplified text is undeniable, paving the way for more
efective simplification systems.</p>
      <p>
        The SimpleText lab [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] is part of CLEF, and it aims to address the task of text simplification of scientific
text. Task 1 in particular [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] involves investigating the performance of text simplification on both
the sentence-level and document-level. It uses the Cochrane-Auto dataset [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] and uses standard text
simplification metrics such as SARI, BLEU, BERTscore, etc. to evaluate the generated simplified text
against the reference ones.
      </p>
      <p>In this paper, we tackle both sentence-level text simplification as well as document-level using a
tiered approach. For the sentence-level, an LLM first generates a simplification strategy regarding a
sentence, and then the LLM is tasked to perform that strategy to generate a simplified version. For the
document-level, the LLM generates a summary of the text as a whole and then uses that summary as
part of the prompt that guides the simplification process.</p>
      <p>The paper is organized as follows: Section 2 provides a small literature review of the text simplification
task; section 3 provides details of the approach undertaken for each of the subtasks; section 4 showcases
the results and provides a brief discussion; section 5 discusses possible future work and the conclusions
we have derived from these experiments.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>Scientific text simplification aims to enhance the accessibility and comprehensibility of technical content
for non-expert audiences, including patients, educators, and the general public. This task has been
explored at both sentence and document levels, with increasing interest in using neural and large
language model (LLM)-based methods.</p>
      <p>
        Ondov et al. [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] provide a comprehensive survey of automated methods for biomedical text
simplification. They categorize approaches into rule-based, statistical, and neural systems, highlighting
the trade-ofs between linguistic control and generative fluency. Their analysis underscores
challenges in maintaining factual consistency and domain-specific accuracy, particularly in biomedical
domains. This motivates the need for more grounded, interpretable approaches, such as plan-driven or
summary-guided simplification, which we explore in this paper.
      </p>
      <p>
        Recent work has also examined the role of LLMs in improving user comprehension and reducing
cognitive load. Guidroz et al. [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] evaluate how LLM-based simplification afects the understanding of the
readers and the mental efort of diferent audiences. They find that while LLM-generated simplifications
generally improve readability, there is a risk of hallucination and over-simplification, especially in
scientific and biomedical texts.
      </p>
      <p>
        Fang et al. [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] propose a hierarchical strategy involving LLMs to simplify document level sentences
using a progressive process, that breaks it down to discourse-level, topic-level, and lexical-level
simplification. The task is formulated as a conditional generation problem by autoregressively conditioning
on the input source document. This approach efectively preserves the content of the document while
eliminating ambiguity and subjectivity, and avoids treating the document simplification task as merely
document summarization.
      </p>
      <p>These findings support our motivation to investigate structured prompting methods to enhance
control, coherence, and factuality in the simplification of scientific texts.</p>
      <sec id="sec-2-1">
        <title>2.1. Evaluation Metrics</title>
        <p>A variety of automatic evaluation metrics are employed to assess the quality of generated or simplified
text. In this work, we consider four commonly used metrics: SARI, BLEU, BERTScore (F1) and
FleschKincaid grade level (FKGL), each capturing diferent aspects of text quality to compare LLM-based
simplification strategies.</p>
        <p>
          SARI (System output Against References and against the Input sentence) [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] is specifically designed
for the text simplification task. Unlike traditional metrics, SARI compares the system output not only
to reference simplifications but also to the original input. It evaluates the quality of three operations:
keeping relevant words, deleting unnecessary ones, and adding appropriate new content. SARI is
computed as the average of F1 scores for these three operations and is typically scaled from 0 to 100,
with higher scores indicating better simplification quality.
        </p>
        <p>
          BLEU (Bilingual Evaluation Understudy) [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ] is an n-gram precision-based metric widely used in
machine translation and text generation. It measures the overlap between system outputs and reference
texts, incorporating a brevity penalty to discourage overly short outputs. However, BLEU is less suitable
for simplification, as it tends to penalize edits that diverge lexically from the reference even when such
changes improve simplicity or meaning.
        </p>
        <p>
          BERTScore [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ] evaluates the semantic similarity between the generated text and the reference using
contextual embeddings from a pretrained transformer model. The F1 variant computes the harmonic
mean of precision and recall based on cosine similarity between token embeddings. BERTScore is
particularly useful in capturing semantic adequacy, especially when lexical overlap is low but the
meaning is preserved.
        </p>
        <p>
          Flesch–Kincaid Grade Level (FKGL) [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ] is a readability metric that estimates the U.S. school
grade level required to comprehend the text. It is computed using average sentence length and average
syllables per word. Lower FKGL scores indicate simpler text and are often used as a proxy for evaluating
readability in simplification tasks. However, FKGL focuses solely on surface-level features and does not
consider syntactic or semantic correctness [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ].
        </p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Methodology</title>
      <p>The focus of task 1 is to study the performance of simplification systems in both sentence-level and
document-level settings.</p>
      <sec id="sec-3-1">
        <title>3.1. Sentence-level simplification - Task 1.1</title>
        <p>
          Inspired by recent advances in plan-driven sentence simplification [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ], we adopt a large language model
(llama-3.3-70b-versatile) as a plan-based simplifier. As illustrated in Figure 1, this approach
utilizes few-shot prompting with three inputs: a complex sentence, its corresponding source document,
and the next complex sentence from the document.
        </p>
        <p>The task is structured into two stages. In the first stage, the model is prompted to select an appropriate
simplification strategy from a predefined set: rephrase, delete, split, ignore, or merge. In the second
stage, the model is prompted to generate the corresponding simplified sentence based on the selected
strategy[Appendix B.1].</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Document-level simplification - Task 1.2</title>
        <p>
          In this task, we leverage large language models (LLMs) for summary-guided document simplification [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ].
Figure 2 outlines our two-step pipeline, where a large language model, llama-3.3-70b-versatile,
is employed both as a summarizer and a simplifier.
        </p>
        <p>First, the model is prompted to produce a clear and concise summary of the input complex
document[Appendix B.2]. This summary serves as a semantic scafold to guide the simplification process. In
the second step, the same model is prompted to simplify the original document using the generated
summary as contextual guidance[Appendix B.3]. This strategy enhances coherence and faithfulness
while reducing the risk of over-simplification.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Results</title>
      <sec id="sec-4-1">
        <title>4.1. Evaluation of Task 1.1: Sentence-level Scientific Text Simplification</title>
        <p>The plan_guided_llama system demonstrates efective sentence-level simplification on the 37 aligned
Cochrane-auto abstracts, achieving a strong SARI score of 42.33 and reducing lexical complexity
(8.52) while maintaining reasonable BLEU and FKGL values. This indicates that the model balances
simplification and content preservation. Detailed results are shown in Table 1.
e
r
o
c
S
y
t
i
x
e
l
p
m
o
C
l
a
c
i
x
e</p>
        <p>L
8.89
8.71
8.52
Method</p>
        <p>The plan_guided_llama model demonstrates strong simplification capability on the 217 Plain
Language Summaries test set, achieving a SARI score of 42.98, which reflects balanced simplification
performance. However, its BLEU score (6.33) is notably low, suggesting limited surface-level overlap
with references. The model produces simplified outputs with a significantly lower FKGL (7.82) compared
to the source (13.29), indicating improved readability. The compression ratio (0.48) and low exact copy
rate (0.00) suggest aggressive simplification, while the deletions proportion (0.71) is higher than additions
(0.18), pointing to a tendency to simplify by removal. Full results are shown in Table 2.</p>
        <p>Overall, the results demonstrate that the plan-guided LLaMA system efectively simplifies complex
biomedical text at sentence-level while maintaining readability and informativeness, with a trade-of in
exact lexical overlap.
4.2. Evaluation of Task 1.2: Document-level Scientific Text Simplification
The results in Table 3 show that the llama_summary_simplification system achieves a SARI score
of 40.32, indicating moderate simplification quality. While the BLEU score (7.63) is comparatively
low—suggesting some divergence from reference phrasing—the FKGL score of 9.56 reflects improved
readability relative to the source. The compression ratio of 0.59 and deletion proportion of 0.70 suggest
aggressive content reduction, contributing to simplification but potentially at the cost of semantic
ifdelity.
1.00
0.00
0.00
n
o
i
t
r
o
p
o
r
P
s
n
o
i
t
i
d
d</p>
        <p>A
0.00
0.16
0.31
0.00
0.60
0.70
e
r
o
c
S
y
t
i
x
e
l
p
m
o
C
l
a
c
i
x
e</p>
        <p>L
9.05
8.65
8.50
e
r
o
c
S
y
t
i
x
e
l
p
m
o
C
l
a
c
i
x
e</p>
        <p>L
8.89
8.71
8.49
Method</p>
        <p>The llama_summary_simplification method demonstrates efective simplification on the 217
Plain Language Summaries, achieving a strong SARI score of 42.92 and a reduced FKGL of 9.94. The
lexical complexity score of 8.55 indicates simplification in vocabulary. However, a relatively low BLEU
score of 5.32 and Levenshtein similarity of 0.39 suggest reduced semantic similarity to the reference
(Table 4).</p>
        <p>Overall, the summary guided method achieves consistent simplification across both the datasets,
balancing lower complexity with reduced semantic fidelity.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.3. Comparing LLM based Sentence Simplification Strategies</title>
      </sec>
      <sec id="sec-4-3">
        <title>4.4. Comparing LLM based Document Simplification Strategies</title>
        <p>1.00
0.00
0.00
n
o
i
t
r
o
p
o
r
P
s
n
o
i
t
i
d
d</p>
        <p>A
0.00
0.29
0.24
n
o
i
t
r
o
p
o
r
P
s
n
o
i
t
e
l
e</p>
        <p>D
0.00
0.63
0.75
e
r
o
c
S
y
t
i
x
e
l
p
m
o
C
l
a
c
i
x
e</p>
        <p>L
9.05
8.65
8.55
While the direct LLM-based simplification approach slightly outperforms the summary-guided
method in terms of standard metrics such as SARI, BLEU, and BERTScore, the summary-guided
approach ofers distinct benefits. By first generating a high-level summary and then using it to steer the
simplification process, this method can produce more coherent and purpose-driven simplifications. The
separation of summarization and simplification stages allows the model to better focus on key ideas,
potentially reducing redundancy and irrelevant elaborations.</p>
        <p>Moreover, the reduced token length in the summary-guided method suggests more concise output,
which can be beneficial in applications where brevity and focus are important. Although the readability
score (FKGL) is slightly higher, indicating marginally more complex language, the summary-guided
approach may improve factual alignment and structural cohesion.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusions</title>
      <p>This study explores the efectiveness of large language models (LLMs) in text simplification across
both sentence and document levels. At the sentence level, our plan-driven approach, which prompts
the model to select an explicit simplification strategy before generation, yields improved performance
in fluency and readability compared to direct simplification. At the document level, we propose a
summary-guided simplification pipeline that, while slightly underperforming in standard metrics, ofers
qualitative advantages in conciseness and coherence by leveraging intermediate summarization as
contextual scafolding.</p>
      <p>Our work demonstrates that incorporating structural planning and summarization can enhance
LLM-based simplification, especially for longer and more complex texts. Future research will focus on
1Only SARI is the oficial evaluation metric for Task 1. Other metrics are reported for supplementary analysis.
improving the structural prompting framework by introducing an iterative loop that refines prompts
based on automatic evaluation metrics, serving as a built-in feedback mechanism.</p>
    </sec>
    <sec id="sec-6">
      <title>Acknowledgements</title>
      <p>
        We thank the Data Science at Georgia Tech (DS@GT) CLEF competition group for their support. This
research was supported in part through research cyberinfrastructure resources and services provided by
the Partnership for an Advanced Computing Environment (PACE) at the Georgia Institute of Technology,
Atlanta, Georgia, USA [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ].
      </p>
    </sec>
    <sec id="sec-7">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the authors used ChatGPT and Gemini for grammar and spelling
check, as well as assistance in the code for the conducted experiments. After using these tools, the
authors reviewed and edited the content as needed and take full responsibility for the publication’s
content.</p>
    </sec>
    <sec id="sec-8">
      <title>A. Codabench Competition Submissions</title>
      <sec id="sec-8-1">
        <title>A.1. Submissions for Task 1.1</title>
      </sec>
      <sec id="sec-8-2">
        <title>A.2. Submissions for Task 1.2</title>
      </sec>
    </sec>
    <sec id="sec-9">
      <title>B. Prompt Templates</title>
      <sec id="sec-9-1">
        <title>B.1. LLM Prompt for Plan-Driven Sentence Simplification</title>
        <p>You a r e a s e n t e n c e s i m p l i f i e r .</p>
        <p>Given a document , a s e n t e n c e from t h a t document , and t h e n e x t
s e n t e n c e f o r c o n t e x t , c h o o s e an i n t e r n a l s i m p l i f i c a t i o n s t r a t e g y
from t h e f o l l o w i n g o p t i o n s :
’ r e p h r a s e ’ , ’ d e l e t e ’ , ’ s p l i t ’ , ’ i g n o r e ’ , ’ merge ’ .</p>
        <p>Then o u t p u t ONLY t h e s i m p l i f i e d s e n t e n c e , b a s e d on yo u r c h o s e n
s t r a t e g y .</p>
        <p>Document : The e c o n o m i c r e p o r t showed a s i g n i f i c a n t downturn i n t h e
l a s t q u a r t e r .</p>
      </sec>
      <sec id="sec-9-2">
        <title>B.2. LLM Prompt for Document Summarization</title>
        <p>You a r e g i v e n a complex document . Your t a s k i s t o w r i t e a c l e a r and
c o n c i s e summary t h a t c a p t u r e s t h e e s s e n t i a l i n f o r m a t i o n , main
arguments , and key f i n d i n g s .</p>
        <p>G u i d e l i n e s :
− Do not i n c l u d e minor d e t a i l s or examples u n l e s s c r u c i a l t o t h e
main i d e a .
− Focus on t h e o v e r a l l message and s t r u c t u r e o f t h e document .
− Use s i m p l e and a c c e s s i b l e l a n g u a g e .
− The summary s h o u l d be u n d e r s t a n d a b l e w i t h o u t r e a d i n g t h e o r i g i n a l
document .
### Document :
{ document }
### Summary :</p>
      </sec>
      <sec id="sec-9-3">
        <title>B.3. LLM Prompt for Summary-Guided Document Simplification</title>
        <p>Listing 1: Prompt for LLM-Based Summary-Guided Document Simplification
You a r e g i v e n a complex document and i t s summary . Your t a s k i s t o
r e w r i t e t h e complex document i n a s i m p l e r , c l e a r e r way w h i l e
e n s u r i n g t h e meaning a l i g n s with t h e p r o v i d e d summary .
G u i d e l i n e s :
− Keep t h e r e w r i t t e n v e r s i o n f a i t h f u l t o both t h e o r i g i n a l document
and i t s summary .
− Use simple , a c c e s s i b l e v o c a b u l a r y and s e n t e n c e s t r u c t u r e s .
− Avoid i n t r o d u c i n g new i n f o r m a t i o n not p r e s e n t i n t h e o r i g i n a l
document .
− R e t a i n t h e key i d e a s , s t r u c t u r e , and i n t e n t c a p t u r e d i n t h e
summary .
### Complex Document :
{ document }
### Summary :
{ summary }
### S i m p l i f i e d Document :</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>H.</given-names>
            <surname>Saggion</surname>
          </string-name>
          , Automatic Text Simplification, Springer International Publishing, Cham,
          <year>2022</year>
          . URL: https://link.springer.com/book/10.1007/978-3-
          <fpage>031</fpage>
          -02166-4. doi:
          <volume>10</volume>
          .1007/978-3-
          <fpage>031</fpage>
          -02166-4.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>L.</given-names>
            <surname>Ermakova</surname>
          </string-name>
          , et al.,
          <article-title>Overview of clef 2025 simpletext track: Simplify scientific texts (and nothing more)</article-title>
          , in: J.
          <string-name>
            <surname>Carillo de Albornoz</surname>
          </string-name>
          , et al. (Eds.),
          <source>Experimental IR Meets Multilinguality, Multimodality, and Interaction. Proceedings of the Fifteenth International Conference of the CLEF Association (CLEF</source>
          <year>2025</year>
          ), LNCS, Springer-Verlag,
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>J.</given-names>
            <surname>Bakker</surname>
          </string-name>
          , et al.,
          <article-title>Overview of the clef 2025 simpletext task 1: Simplify scientific text</article-title>
          , in: G.
          <string-name>
            <surname>Faggioli</surname>
          </string-name>
          , et al. (Eds.),
          <source>Working Notes of the Conference and Labs of the Evaluation Forum (CLEF</source>
          <year>2025</year>
          ), CEUR Workshop Proceedings, CEUR-WS.org,
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>J.</given-names>
            <surname>Bakker</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Kamps</surname>
          </string-name>
          ,
          <article-title>Cochrane-auto: An aligned dataset for the simplification of biomedical abstracts</article-title>
          , in: M.
          <string-name>
            <surname>Shardlow</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Saggion</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Alva-Manchego</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Zampieri</surname>
          </string-name>
          , K. North, S. Štajner, R. Stodden (Eds.),
          <source>Proceedings of the Third Workshop on Text Simplification, Accessibility and Readability (TSAR</source>
          <year>2024</year>
          ),
          <article-title>Association for Computational Linguistics</article-title>
          , Miami, Florida, USA,
          <year>2024</year>
          , pp.
          <fpage>41</fpage>
          -
          <lpage>51</lpage>
          . URL: https://aclanthology.org/
          <year>2024</year>
          .tsar-
          <volume>1</volume>
          .5/. doi:
          <volume>10</volume>
          .18653/v1/
          <year>2024</year>
          .tsar-
          <volume>1</volume>
          .5.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>B.</given-names>
            <surname>Ondov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Attal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Demner-Fushman</surname>
          </string-name>
          ,
          <article-title>A survey of automated methods for biomedical text simplification</article-title>
          ,
          <source>Journal of the American Medical Informatics Association</source>
          <volume>29</volume>
          (
          <year>2022</year>
          )
          <fpage>1976</fpage>
          -
          <lpage>1988</lpage>
          . doi:
          <volume>10</volume>
          .1093/jamia/ocac149.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>T.</given-names>
            <surname>Guidroz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Ardila</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mansour</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Jhun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Gonzalez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Ji</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Sanchez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Kakarmath</surname>
          </string-name>
          ,
          <string-name>
            <surname>M. M. Bellaiche</surname>
            ,
            <given-names>M. Ángel</given-names>
          </string-name>
          <string-name>
            <surname>Garrido</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Ahmed</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Choudhary</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Hartford</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Xu</surname>
            ,
            <given-names>H. J. S.</given-names>
          </string-name>
          <string-name>
            <surname>Echeverria</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Shafer</surname>
            , Eric, Cao,
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>Matias</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Hassidim</surname>
            ,
            <given-names>D. R.</given-names>
          </string-name>
          <string-name>
            <surname>Webster</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>Liu</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Fujiwara</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Bui</surname>
            ,
            <given-names>Q.</given-names>
          </string-name>
          <string-name>
            <surname>Duong</surname>
          </string-name>
          ,
          <article-title>Llm-based text simplification and its efect on user comprehension and cognitive load</article-title>
          ,
          <year>2025</year>
          . arXiv:
          <fpage>2505</fpage>
          .
          <year>01980</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>D.</given-names>
            <surname>Fang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Qiang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Yuan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <article-title>Progressive document-level text simplification via large language models</article-title>
          ,
          <year>2025</year>
          . URL: https://arxiv.org/abs/2501.03857. arXiv:
          <volume>2501</volume>
          .03857, preprint.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>W.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Napoles</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Pavlick</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Callison-Burch</surname>
          </string-name>
          ,
          <article-title>Optimizing statistical machine translation for text simplification</article-title>
          ,
          <source>Transactions of the Association for Computational Linguistics</source>
          <volume>4</volume>
          (
          <year>2016</year>
          )
          <fpage>401</fpage>
          -
          <lpage>415</lpage>
          . URL: https://aclanthology.org/Q16-1029/. doi:
          <volume>10</volume>
          .1162/tacl_a_
          <fpage>00107</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>K.</given-names>
            <surname>Papineni</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Roukos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Ward</surname>
          </string-name>
          , W.-J. Zhu,
          <article-title>Bleu: a method for automatic evaluation of machine translation, in: Proceedings of the 40th Annual Meeting of the Association for Computational Linguistics (ACL), Association for Computational Linguistics</article-title>
          , Philadelphia, PA, USA,
          <year>2002</year>
          , pp.
          <fpage>311</fpage>
          -
          <lpage>318</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>T.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Kishore</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. Q.</given-names>
            <surname>Weinberger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Artzi</surname>
          </string-name>
          , Bertscore:
          <article-title>Evaluating text generation with bert</article-title>
          ,
          <source>in: International Conference on Learning Representations</source>
          ,
          <year>2020</year>
          . ArXiv preprint arXiv:
          <year>1904</year>
          .09675.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>J. P.</given-names>
            <surname>Kincaid</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. P.</given-names>
            <surname>Fishburne</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. L.</given-names>
            <surname>Rogers</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B. S.</given-names>
            <surname>Chissom</surname>
          </string-name>
          ,
          <source>Derivation of New Readability Formulas (Automated Readability Index, Fog Count and Flesch Reading Ease Formula) for Navy Enlisted Personnel, Research Branch Report 8-75</source>
          , Naval Technical Training, U.S. Naval Air Station, Memphis,
          <string-name>
            <surname>TN</surname>
          </string-name>
          ,
          <year>1975</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>T.</given-names>
            <surname>Tanprasert</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Kauchak</surname>
          </string-name>
          ,
          <article-title>Flesch-kincaid is not a text simplification evaluation metric</article-title>
          , in: A.
          <string-name>
            <surname>Bosselut</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          <string-name>
            <surname>Durmus</surname>
            ,
            <given-names>V. P.</given-names>
          </string-name>
          <string-name>
            <surname>Gangal</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Gehrmann</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>Jernite</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Perez-Beltrachini</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Shaikh</surname>
          </string-name>
          , W. Xu (Eds.),
          <source>Proceedings of the 1st Workshop on Natural Language Generation</source>
          , Evaluation, and
          <string-name>
            <surname>Metrics</surname>
          </string-name>
          (GEM
          <year>2021</year>
          ),
          <article-title>Association for Computational Linguistics</article-title>
          , Online,
          <year>2021</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>14</lpage>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <year>2021</year>
          .gem-
          <volume>1</volume>
          .1.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <surname>PACE</surname>
          </string-name>
          ,
          <article-title>Partnership for an Advanced Computing Environment (PACE</article-title>
          ),
          <year>2017</year>
          . URL: http://www. pace.gatech.edu.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>