<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Text Adaptation to Plain Language and Easy Read via Automatic Post-Editing Cycles</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Jesús Calleja</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>David Ponce</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Thierry Etchegoyhen</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Fundación Vicomtech, Basque Research and Technology Alliance (BRTA)</institution>
          ,
          <addr-line>Donostia - San Sebastián</addr-line>
          ,
          <country country="ES">Spain</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>University of the Basque Country - UPV/EHU</institution>
          ,
          <addr-line>Donostia - San Sebastián</addr-line>
          ,
          <country country="ES">Spain</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <abstract>
        <p>We describe Vicomtech's participation in the CLEARS challenge on text adaptation to Plain Language and Easy Read in Spanish. Our approach features automatic post-editing of diferent types of initial Large Language Model adaptations, where successive adaptations are generated iteratively until readability and similarity metrics indicate that no further adaptation refinement can be successfully performed. Taking the average of all oficial metrics, our submissions achieved first and second place in Plain language and Easy Read adaptation, respectively.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Plain language</kwd>
        <kwd>Easy Read</kwd>
        <kwd>Text Simplification</kwd>
        <kwd>Text Adaptation</kwd>
        <kwd>Automatic Post-editing</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Text adaptation to simplified language varieties is an essential task to provide accessible content for large
segments of the population, including people with reading dificulties, cognitive disabilities, learning
dificulties, or low literacy in general [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. Plain Language (PL) and Easy Read (ER) are two variants of
adaptation with specific characteristics, partly determined by the specific audience they target, although
both share a need for simplified grammar and vocabulary. Whereas PL focuses on communicating
information via clear and concise language [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], ER includes specific requirements for its target audience
in terms of sentence length, explanation of complex terms, or text segmentation, among others [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ].
      </p>
      <p>
        The CLEARS challenge, organised within the IberLEF 2025 conference [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], aimed to assess the quality
of competing approaches to both PL and ER adaptation in Spanish, each type of adaptation being
addressed as a separate task. In each case, a set of input documents was provided, along with their
respective human adaptation. Participants had to provide their own automated adaptations, which were
measured with a combination of three metrics: embedding (EMB) and bag of words (BoW) similarities
between the proposed adaptation and human references, and the Fernandez Huerta readability index
[
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] for the proposed adaptation. Further details on the CLEARS shared task are provided in Botella-Gil
et al. [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ].
      </p>
      <p>In this work, we describe Vicomtech’s participation in the shared task, detailing our specific approach
and results. We tackled both types of adaptation within a unique framework, based on Automatic
Post-Editing Cycles (APEC). We leverage Large Language Models (LLM) to provide initial adaptations
via either Retrieval-Augmented Generation (RAG) or model tuning with Direct Preference Optimisation
(DPO). These initial adaptations are further refined via APEC, focusing the adaptation on guideline
compliance, text readability, and information preservation.</p>
      <p>Taking the average of all oficial metrics, our approach was the top-performing system for PL
adaptation and second-best for ER adaptation. In the remainder of this paper, we discuss related work,
the APEC approach and our results on both development and test sets. We conclude with future lines
of research based on the current limitations of the explored approaches and evaluation setups.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>
        Text Simplification. Automatic Text Simplification (ATS) aims to enhance text accessibility by
reducing linguistic complexity while preserving meaning [
        <xref ref-type="bibr" rid="ref1 ref7">1, 7</xref>
        ] and has a long standing tradition in
natural language processing [
        <xref ref-type="bibr" rid="ref10 ref8 ref9">8, 9, 10</xref>
        ]. The task involves simplification along diferent dimensions,
notably lexical simplification [
        <xref ref-type="bibr" rid="ref11 ref12 ref13">11, 12, 13</xref>
        ] and syntactic rephrasing [
        <xref ref-type="bibr" rid="ref14 ref15 ref16">14, 15, 16</xref>
        ].
      </p>
      <p>
        ATS has been performed via diferent approaches over the years. Starting with rule-based
simplification [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ], the field moved towards data-driven approaches that exploited aligned pairs of complex
and simplified sentences [
        <xref ref-type="bibr" rid="ref18">18, 19</xref>
        ]. End-to-end ATS modelling has been the focus of recent work in the
ifeld, most based on variants of the Transformer neural architecture [ 20] adapted for the task. Thus,
control parameters have been successfully employed to guide simplification, such as sentence length,
Levenshtein distance, or dependency graph depth [21, 22, 23, 24]. Generative Large Language Models
(LLM) have been at the forefront of recent ATS research, following the trend in most downstream
natural language processing tasks [25, 26, 27, 28, 29].
      </p>
      <p>Related to our APEC approach is the work of Guidroz et al. [30], whose method includes iterative
simplification via prompt refinement and automated evaluation. Although their approach difers in
several respects, mainly prompt refinement and a diferent set of evaluation criteria, both approaches
share an iterative process to for LLM-based adaptation to simpler language.1
Plain Language. Plain Language refers to a writing style that is clear, direct, and accessible to a
broad audience. It has been normalised into an international standard (ISO 24495-1:2023),2 with notable
adoption eforts in multiple countries. 3 The standard emphasises the use of common words, avoiding
technical terms as much as possible, along with the use of short sentences</p>
      <p>These include common, everyday words, except for necessary technical terms. Other qualities include
the use of personal pronouns; the active voice; logical organization; and easy-to-read and understandable
design features, such as bullets and tables.</p>
      <p>
        PL guidelines are increasingly adopted as constraints for simplification systems. For instance, within
the CLARA-NLP project [31], PL criteria have been integrated into text processing and simplification
pipelines for public service documents. The CLEARS challenge is another step towards the development
and evaluation of automated adaptation to Plain Language [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ].
      </p>
      <p>Easy Read. Easy Read, also referred to as Easy-to-read, is a specialised form of text adaptation that
targets individuals with intellectual disabilities, reading dificulties, or low functional literacy. A set of
guidelines have been established over the years, in particular from Inclusion Europe,4 and a standard was
defined for the Spanish language (NE 153101:2018 EX). 5 ER recommendations include the use of short
and simple sentences, frequent vocabulary, and explicit explanation of dificult concepts, among others.
The use of supportive visual elements and structured formatting, including sentence segmentation, are
also part of the general recommendations.</p>
      <p>Easy Read presents unique challenges for existing ATS systems. It demands simplifications that go
beyond surface-level lexical substitution, often requiring sentences splitting and rewriting, omission
of implicit information, and reformulation of abstract terms. Although there is still a relative lack of
empirical research overall in the field of ER [ 32], there is an increasing trend towards automated ER
adaptation. Thus, recent models have targeted ER constraints either individually, e.g., text segmentation
[33], or holistically via end-to-end adaptation using prompt engineering and fine-tuning over ER data
[34]. New practical applications that integrate advances in the field of automatic ER adaptation have
also been proposed recently, in particular for Spanish [35, 36, 37].
1Note that our approach was developed independently for the challenge, prior to the publication of their work.
2See footnote 1.
3https://plainlanguagenetwork.org/plain-language/plain-language-around-the-world/
4See footnote 2.
5https://www.une.org/encuentra-tu-norma/busca-tu-norma/norma?c=N0060036</p>
    </sec>
    <sec id="sec-3">
      <title>3. Models and Tools</title>
      <p>For both corpus analysis and adaptation experiments, we used a combination of in-house scripts and
publicly available models and tools. The shared task involved two main types of metrics:6 (i) cosine
similarity between the generated text and human adaptation references, and (ii) the Fernández Huerta
Readability Index, a metric based on the Flesch-Kincaid formula, adapted for Spanish, which measures
readability in terms of average sentence and syllable length. Cosine similarity was in turn split into two
metrics:7 bag-of-words similarity, measuring lexical overlap, and text embedding similarity.</p>
      <p>For cosine similarity of text embeddings, we used the paraphrase-multilingual-mpnet-base-v2 model
from the SentenceTransformers library[38].8 For BoW similarity, we computed the cosine similarity
between bag-of-words representations generated with the CountVectorizer of the ScikitLearn library
[39].9 The text was processed with the spaCy library [40],10 using model es_core_news_sm. FH scores
were computed with an external script: https://github.com/alejandromunozes/legibilidad.</p>
      <sec id="sec-3-1">
        <title>All of our experiments were conducted with the LLaMA 3.1 Instruct 8B model [41].11</title>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Corpora</title>
      <p>The ClearSim corpus provided for the CLEARS challenge initially consisted of 2,400 original documents
per task, each accompanied by corresponding PL and ER adaptations [42, 43]. To accommodate context
limitations within our experimental setup, we applied a filtering process to remove instances where the
combined token count of the instruction, original input, and adapted output, exceeded 3,000 tokens.
Tokenisation was performed using the LLaMA 3.1 Instruct 8B tokeniser. Following this filtering step,
2,348 and 2,347 samples remained for the PL and ER tasks, respectively.</p>
      <p>We partitioned the data by allocating 240 documents to a development set (dev set), with the remainder
used for training. Detailed statistics for each partition across both tasks are presented in Tables 1 and 2.
We also indicate, as a reference, the BoW and EMB similarity scores between reference adaptation and
input text, as well as the FH index.
6https://sites.google.com/gcloud.ua.es/clears/evaluation-metrics
7https://www.kaggle.com/competitions/clears-adaptation-texts
8https://huggingface.co/sentence-transformers/paraphrase-multilingual-mpnet-base-v2
9https://scikit-learn.org
10https://spacy.io/
11https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct
texts by 6.07 and 6.50 points for PL and ER adaptation, respectively, on the train set; for the dev set, the
diferences amounted to 5.20 and 6.88 points, respectively. Average similarity was slightly higher on
the dev set, although both tasks featured a relatively low similarity between input and adapted texts,
with at most 0.71 similarity for the PL dev set. This is mainly due to low BoW similarity, which is
expected considering the lexical changes performed during PL or ER adaptation. Embedding similarity
was comparatively higher, in the [0.86, 0.88] range, indicating a relatively high degree of information
preservation between original and adapted texts.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Automatic Adaptation via Post-Editing Cycles</title>
      <p>Our approach consists of two main phases: initial adaptation, followed by refinement via automatic
post-editing cycles. We describe each step in turn below.
5.1. Initial Adaptation
The initial adaptations were performed with specific LLM prompts for the PL and ER tasks, adapted
from publicly available guidelines.12 We explored the following variants for both tasks:
• Zero-shot (ZS): direct querying of the model with the selected prompts.
• Few-shot (FS): complementing the prompts with five selected demonstrations sampled from the
training set. For this approach, we experimented with the following variants:
– FS RDM: random demonstration sampling.
– FS BM25: lexical RAG with BM25 indexing and querying.</p>
      <p>– FS SIM: semantic similarity RAG with embedding indexing and querying.
• Fine-tuning (FT): supervised fine-tuning over the training data.
• Direct Preference Optimisation (DPO) : direct preference optimisation [44] over contrastive
pairs, where the references from the training data are taken as the preferred output over
corresponding zero-shot generation on the same input.</p>
      <p>BM25 and embedding-based retrieval were performed with in-house scripts, using the five most
similar samples throughout our experiments. For BM25, we leveraged the BM25Okapi library [45],
using a minimum length of four characters for indexing and white-space tokenisation.13 For
embedding similarity, we used the LangChain library,14 the SemanticSimilarityExampleSelector class, and
12See Appendix A for details.
13https://github.com/dorianbrown/rank_bm25
14https://github.com/langchain-ai/langchain
the paraphrase-multilingual-mpnet-base-v2 model (op. cit.). For the FS BM25 approach, we further
experimented with filtering candidate demonstrations based on length ratio constraints, excluding
adaptations whose length ratio (adaptation length relative to the source) fell outside predefined ranges.
Specifically, we tested threshold ranges of [0.8, 1.2] and [0.5, 1.5].</p>
      <p>Generation was performed with the Hugginface pipeline, using a default greedy decoding strategy
with fixed hyperparameters: do_sample = False and num_beams=1. For the DPO approach, we further
experimented with sampling based on the following values: temperature = 0,3; top_k = 40; p=0.95,</p>
      <sec id="sec-5-1">
        <title>For fine-tuning, we employed a QLoRA approach [ 46] using the SFT Trainer15 from Hugging Face,</title>
        <p>using parameters: dropout=0.05, alpha=8, r=16. We targeted the query and value modules, used a
learning rate of 3e-4, and the model was loaded in 4 bits. All other parameters were per their default
values. Models were trained until convergence, with an early stopping patience set to 5. Finally, DPO
was performed with the HugginFace TRL library,16 using default values.</p>
        <p>The comparative results for the initial adaptation phase are shown in Table 3 and Table 4 for Task 1
and Task 2, respectively.</p>
        <p>For PL adaptation, in terms of average between similarity and FH, both FT and DPO achieved higher
scores overall, although few-shot variants also achieved relatively high average scores, fueled by higher
FH readabilty scores and outperforming zero-shot inference. Fine-tuning consistently outperformed
other variants in terms of BoW, as expected from tuning over human training references, at the cost of
lower FH overall. DPO was more balanced in this respect, outperforming FS BM25 in terms of BoW
15https://huggingface.co/docs/trl/sft_trainer
16https://huggingface.co/docs/trl/en/dpo_trainer
but also achieving higher FH scores. FS BM25 achieved balanced results, excepting BoW similarity,
where the lack of tuning over human preferences led to more lexically diverse output. Of note are the
consistently better results achieved by BM25 over embedding similarity retrieval of demonstrations,
an indication of the dificulties in retrieving similar examples based solely on the general semantic
similarity on relatively long input samples.</p>
        <p>For ER adaptation, FS BM25 variants were the more robust alternative in terms of average and
embedding similarity, although DPO significantly outperformed all variants in terms of FH, with
finetuning being markedly worse on this metric. Filtering the sampled data by length range for BM25 led to
minor improvements in terms of FH on both tasks, with lower embedding similarity overall. Since these
diferences were marginal overall, and filtering runs the risk of discarding optimal demonstrations, we
discarded any type of filtering for follow-up experiments.</p>
        <p>Considering these results, for both tasks we selected the BM25 and DPO variants as initial adaptations
to feed the APEC cycles, described in the next section.
5.2. APEC</p>
        <p>The APEC approach is illustrated in Figure 1. Our method aimed to provide gradual adaptation
refinement over initial hypotheses, based on the fact that initial adaptations may sufer from hallucinations
or deviate from adaptation guidelines. Another underlying intuition for our approach lies in the fact
that LLMs have demonstrated strong potential in tasks that involve text improvement and reflection
over an input, for instance via Chain of Thought reasoning [47].</p>
        <p>The input to the APEC cycles are an input text and its corresponding adaptation. The latter is
initialised with any preliminary adaptation, which, in our experiments, were either the BM25 or the
DPO adaptation candidates. Based on the input text, we extract few-shot demonstrations from the
training set via BM25 retrieval, to provide further guidance to the post-editing process.17</p>
        <p>The input text, the current adaptation, and the demonstrations, are fed as input into the selected</p>
      </sec>
      <sec id="sec-5-2">
        <title>LLM, whose prompt instructs the model to revise the adaptation based on the task guidelines.18 Based</title>
        <p>on the provided information, the system is tasked to perform separate actions: (i) Analyse the current
adaptation on the basis of the guidelines and demonstrations; (ii) Provide a new adaptation fixing
any identified issues; (iii) Add any final thoughts on the process separately, if needed. In this APEC
approach, the LLM component is thus tasked to act both as a judge of the quality of a given adaptation,
and as a post-editing agent given a set of adaptation criteria.</p>
        <p>The output generated in this step of the process is then evaluated with two metrics: embedding
cosine similarity between the input text and the refined adaptation, to measure meaning preservation,
and the FH index over the generated text to assess readability. The average of both metrics is then
compared with that of the previous adaptation: if the average improved, the input adaptation is replaced
with the refined version; otherwise, the current input is maintained.
17We used 5 demonstrations in all experiments.
18See Appendix A for the complete prompts.</p>
        <p>The iterations are repeated for a fixed number of cycles, 5 in our experiments. The reason not to end
the cycles if no improvement was achieved at a given point is to allow the model to find alternative
adaptations which may still improve over the current one.19</p>
        <p>Note that, in the APEC phase there are no references over which to compute BoW similarity or any
measure of similarity with a human reference. The only actionable references in the process are the
input text and previous adaptations.</p>
        <p>The results over the development set in both tasks are shown in Table 5. We indicate the three metric
scores (FH, BoW similarity and Embedding similarity), their average (FH-Sim avg, and the FH gain over
the original text. In both tasks, the tendencies in terms of FH were similar, with significant gains from
initial adaptations, with BM25 or DPO, to the APEC refined adaptations. These gains, combined with
either minor degradation in terms of embedding similarity for Task 1 or more pronounced losses on the
same metric for Task 2, led to consistently higher average scores for APEC adaptation on both tasks.</p>
        <p>BoW similarity was consistently impacted by APEC, except for DPO in Task 2, indicating that the
cyclic refinements led to adaptations that difered significantly from the initial BM25 or DPO input
adaptations, which are closer to training reference adaptation. This is also an expected side-efect of our
approach, as the refinement process is mainly guided by improving readability with shorter sentences
and words, while maintaining overall semantic text similarity as much as possible.</p>
        <p>Finally, applying APEC over either DPO or BM25 led to relatively similar results overall, showcasing
the flexibility of the refinement process to diferent types of initialisation. In Appendix B, we provide
examples of refinement from initial to final adaptations.</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>6. Task Results</title>
      <p>The statistics for the test data are shown in Table 6. To compare with the data in the training sets, we
computed similar statistics over the source texts.20</p>
      <p>For both tasks, the test sentences were significantly shorter on average: 24.5 words per sentence in
the train set vs. 17.58 for Task 1; 24.19 vs. 17.50 for Task 2. Note that the same contrast exists between
our sampled development data and the source test, with larger sentences in the former as well. In
terms of syllables per word, the test source data also feature notable diferences, with an average of
2.12 for both tasks, compared to 2.07 on the training set. The FH index was also higher for the test
data compared to the training data, and was closer to that of the development data, though the latter
featured higher FH values overall. The source test data thus seems to difer in some respect from the
training data, featuring shorter and relatively simpler text to some degree.</p>
      <p>For our submissions, which were limited to one per task in the challenge, we ensembled the results
achieved via APEC over DPO and BM25, for each task. For each input sample from the test set, we thus
19We use a default temperature of 0.8 to allow for output diversity when the input remained unchanged, which will be the
case whenever a given refinement did not improve over the previous cycle.
20Since the adapted references were not available, similarity metrics could not be computed, naturally.
selected the APEC adaptation which achieved the highest average similarity and readability score. Test
set statistics are shown in Table 6.</p>
      <p>The oficial results are shown in Table 7 and Table 8 for Task 1 and Task 2, respectively. We ordered
the participating systems in the table from top to bottom by taking the average of the (scaled to 100)
similarity and readability scores as unified metric, and indicate the best performing system for each
metric in bold.</p>
      <p>For PL adaptation, our approach was the best-performing system when averaging all oficial metrics
results; for ER adaptation, it achieved second place. APEC adaptation thus proved to be a robust
approach for both types of adaptation overall. Our method outperformed the alternatives in terms of the
readability index in both tasks, showcasing the ability of LLMs to guide the adaptation towards shorter
sentences and words, which are a cornerstone of adaptation towards PL and ER language varieties,
while maintaining overall text similarity.</p>
      <p>In terms of cosine similarity, most approaches achieved comparable results, although it is worth
noting that our adaptation was the most dissimilar for ER adaptation and ranked third out of four
participants for PL adaptation. Our approach mainly sufered in terms of bag of words similarity, as
we ranked last in both tasks on this metric. This is mainly a by-product of our approach, which does
not tune the model towards reference adaptations, beyond the few-shot demonstrations, but rather
towards preserving input information while increasing readability, at the cost of potentially difering
from lexical choices and formulations made in reference adaptations.</p>
    </sec>
    <sec id="sec-7">
      <title>7. Discussion</title>
      <p>Our approach critically relies on automated metrics to assess adaptation progress within our iterative
approach. We employed the metrics selected for the shared task which could be included without
access to references, namely the FH readability index computed over the adapted text, and embedding
similarity between the input and the adaptation hypotheses. Other metrics might be more appropriate
to refine adaptations iteratively, such as the custom evaluation metric of Guidroz et al. [30], or the
reference-less SLE metric proposed by Cripwell et al. [48].</p>
      <p>Assessing the quality of PL or ER adaptation with automated metrics is a complex task, with clear
limitations of typical reference-based metrics [49, 48, 50]. Thus, lexical similarity is bound to measure
closeness to reference adaptation choices, although multiple adaptations might be equally valid. This
issue can be mitigated with multiple references, although their creation is a costly process. Text
embedding similarity also sufers from similar issues, although semantic variation computed in that
manner is less susceptible to lexical choice variation. Metrics such as the FH readability index present
the advantage of being independent of any reference, although their adequacy to properly assess PL or
ER adaptation might be debatable as well [51].</p>
      <p>Human evaluation, by adaptation experts and/or target audiences, would of course be an optimal
option to properly assess the quality of ER or PL adaptation, independently of pre-established references,
although this type of approach comes with a high operational cost. In future work, it might nonetheless
be worth exploring the correlation between the diferent degrees of adaptation featured in our approach
and human evaluations in terms of readability, meaning preservation, and usability for the target
audiences of the diferent types of adaptation.</p>
      <p>Finally, it is worth noting that approaches like APEC come with a higher computational cost than
direct end-to-end methods, as they may require multiple cycles of adaptation refinement in some cases.
They might thus be more suited for ofline text adaptation. For online usage, where users might expect
rapid adaptation, non-iterative adaptation might be a more suitable option, depending on the capacity
of the computational infrastructure. Non-iterative methods, such as BM25 RAG or models tuned via
DPO, which achieve quality results in our experiments, might be alternatives worth considering in this
type of inference scenarios.</p>
    </sec>
    <sec id="sec-8">
      <title>8. Conclusion</title>
      <p>In this work, we described Vicomtech’s participation in the CLEARS challenge, which involved text
adaptation to Plain language and Easy Read language varieties. We first explored a wide range of initial
approaches, based on LLM variants that included zero-shot, few-shot, fine-tuning, and Direct Preference
Optimisation. The optimal variants, namely FS BM25 and DPO, were used as input to our main method,
centred on automatic post-editing cycles (APEC).</p>
      <p>The APEC approach involves iterative refinement of initial adaptations, aiming to increase readability
according to adaptation guidelines, while maintaining semantic similarity with the input text. This
approach improved markedly in terms of the readability index over the initial adaptations via FS BM25
and DPO on the development sets, while maintaining high similarity with the original text.</p>
      <p>In terms of oficial results, averaging the results from all three metrics, our submissions ranked first
on the PL adaptation task, and second on the ER adaptation task. We achieved the highest scores in
terms of the readability index in both tasks, with lower scores for embedding similarity than competing
approaches, although the diferences were rather small between systems. Our approach underperformed
to a larger degree in terms of lexical similarity, which was expected as our APEC approach is not tuned
over reference adaptations, but rather aims to independently generate shorter words, along with other
adaptation characteristics to improve readability. Overall, our results show the viability of the APEC
approach for PL and ER text adaptation.</p>
      <p>In future work, we will explore alternative metrics to both gear adaptation refinements and evaluate
the results, in particular pretrained metrics tuned to adaptation tasks.</p>
    </sec>
    <sec id="sec-9">
      <title>Acknowledgments</title>
      <p>We would like to thank the anonymous reviewers for their helpful comments and suggestions.</p>
    </sec>
    <sec id="sec-10">
      <title>Declaration on Generative AI</title>
      <p>The author(s) have not employed any Generative AI tools.
[19] L. Specia, Translating from complex to simplified sentences, in: International Conference on</p>
      <p>Computational Processing of the Portuguese Language, Springer, 2010, pp. 30–39.
[20] A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, Ł. Kaiser, I. Polosukhin,</p>
      <p>Attention is all you need, Advances in neural information processing systems 30 (2017).
[21] L. Martin, É. V. De La Clergerie, B. Sagot, A. Bordes, Controllable sentence simplification, in:</p>
      <p>Proceedings of the Twelfth Language Resources and Evaluation Conference, 2020, pp. 4689–4698.
[22] L. Martin, A. Fan, É. V. De La Clergerie, A. Bordes, B. Sagot, Muss: Multilingual unsupervised
sentence simplification by mining paraphrases, in: Proceedings of the Thirteenth Language
Resources and Evaluation Conference, 2022, pp. 1651–1664.
[23] K. C. Sheang, H. Saggion, Controllable sentence simplification with a unified text-to-text transfer
transformer, in: A. Belz, A. Fan, E. Reiter, Y. Sripada (Eds.), Proceedings of the 14th International
Conference on Natural Language Generation, Association for Computational Linguistics, Aberdeen,
Scotland, UK, 2021, pp. 341–352. URL: https://aclanthology.org/2021.inlg-1.38/. doi:10.18653/
v1/2021.inlg-1.38.
[24] S. Štajner, K. C. Sheang, H. Saggion, Sentence simplification capabilities of transfer-based models,
in: Proceedings of the AAAI Conference on Artificial Intelligence, volume 36, 2022, pp. 12172–
12180.
[25] P. Laban, T. Schnabel, P. Bennett, M. A. Hearst, Keep it simple: Unsupervised simplification
of multi-paragraph text, in: Proceedings of the 59th Annual Meeting of the Association for
Computational Linguistics and the 11th International Joint Conference on Natural Language
Processing (Volume 1: Long Papers), 2021, pp. 6365–6378.
[26] T. Kew, A. Chi, L. Vásquez-Rodríguez, S. Agrawal, D. Aumiller, F. Alva-Manchego, M. Shardlow,
Bless: Benchmarking large language models on sentence simplification, in: Proceedings of the
2023 Conference on Empirical Methods in Natural Language Processing, 2023, pp. 13291–13309.
[27] Y. Feng, J. Qiang, Y. Li, Y. Yuan, Y. Zhu, Sentence simplification via large language models, arXiv
preprint arXiv:2302.11957 (2023).
[28] S. Agrawal, M. Carpuat, Controlling pre-trained language models for grade-specific text
simplification, in: Proceedings of the 2023 Conference on Empirical Methods in Natural Language
Processing, 2023, pp. 12807–12819.
[29] X. Wu, Y. Arase, An in-depth evaluation of gpt-4 in sentence simplification with error-based
human assessment, arXiv preprint arXiv:2403.04963 (2024).
[30] T. Guidroz, D. Ardila, J. Li, A. Mansour, P. Jhun, N. Gonzalez, X. Ji, M. Sanchez, S. Kakarmath,
M. M. Bellaiche, et al., Llm-based text simplification and its efect on user comprehension and
cognitive load, arXiv preprint arXiv:2505.01980 (2025).
[31] A. Moreno-Sandoval, L. Campillos-Llanos, A. García-Serrano, Language resources in spanish for
automatic text simplification across domains, arXiv preprint arXiv:2409.20466 (2024).
[32] M. González-Sordé, A. Matamala, Empirical evaluation of easy language recommendations: a
systematic literature review from journal research in catalan, english, and spanish, Universal
Access in the Information Society 23 (2024) 1369–1387.
[33] J. Calleja, T. Etchegoyhen, D. Ponce, Automating easy read text segmentation, in: Y. Al-Onaizan,
M. Bansal, Y.-N. Chen (Eds.), Findings of the Association for Computational Linguistics: EMNLP
2024, Association for Computational Linguistics, Miami, Florida, USA, 2024, pp. 11876–11894. URL:
https://aclanthology.org/2024.findings-emnlp.694/. doi: 10.18653/v1/2024.findings-emnlp.
694.
[34] P. Martínez, A. Ramos, L. Moreno, Exploring large language models to generate easy to read content,
Frontiers in Computer Science Volume 6 - 2024 (2024). URL: https://www.frontiersin.org/journals/
computer-science/articles/10.3389/fcomp.2024.1394705. doi:10.3389/fcomp.2024.1394705.
[35] M. C. Suárez-Figueroa, I. Diab, E. Ruckhaus, I. Cano, First steps in the development of a support
application for easy-to-read adaptation, Universal Access in the Information Society 23 (2024)
365–377.
[36] M. Madina, I. Gonzalez-Dios, M. Siegel, Languagetool as a cat tool for easy-to-read in spanish, in:
Proceedings of the 3rd Workshop on Tools and Resources for People with REAding DIficulties
(READI)@ LREC-COLING 2024, 2024, pp. 93–101.
[37] I. Diab, M. C. Suárez-Figueroa, R. Peris, Towards an automatic easy-to-read adaptation of dialogues
in narrative texts in spanish, in: International Conference on Computers Helping People with
Special Needs, Springer, 2024, pp. 208–216.
[38] N. Reimers, I. Gurevych, Sentence-bert: Sentence embeddings using siamese bert-networks,
in: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing,
Association for Computational Linguistics, 2019. URL: http://arxiv.org/abs/1908.10084.
[39] F. Pedregosa, G. Varoquaux, A. Gramfort, V. Michel, B. Thirion, O. Grisel, M. Blondel, P. Prettenhofer,
R. Weiss, V. Dubourg, J. Vanderplas, A. Passos, D. Cournapeau, M. Brucher, M. Perrot, E. Duchesnay,
Scikit-learn: Machine learning in Python, Journal of Machine Learning Research 12 (2011) 2825–
2830.
[40] M. Honnibal, I. Montani, S. Van Landeghem, A. Boyd, spaCy: Industrial-strength Natural Language</p>
      <p>Processing in Python, 2020. doi:10.5281/zenodo.1212303.
[41] A. Grattafiori, A. Dubey, A. Jauhri, A. Pandey, A. Kadian, A. Al-Dahle, A. Letman, A. Mathur,</p>
      <p>A. Schelten, A. Vaughan, et al., The llama 3 herd of models, arXiv preprint arXiv:2407.21783 (2024).
[42] I. Espinosa-Zaragoza, J. Abreu-Salas, P. Moreda, M. Palomar, Automatic text simplification for
people with cognitive disabilities: Resource creation within the ClearText project, in: S. Štajner,
H. Saggio, M. Shardlow, F. Alva-Manchego (Eds.), Proceedings of the Second Workshop on Text
Simplification, Accessibility and Readability, INCOMA Ltd., Shoumen, Bulgaria, Varna, Bulgaria,
2023, pp. 68–77. URL: https://aclanthology.org/2023.tsar-1.7/.
[43] B. Botella-Gil, I. Espinosa-Zaragoza, P. Moreda, M. Palomar, Corpus ClearSim, 2024. URL: http:
//hdl.handle.net/10045/151688.
[44] R. Rafailov, A. Sharma, E. Mitchell, C. D. Manning, S. Ermon, C. Finn, Direct preference
optimization: Your language model is secretly a reward model, Advances in Neural Information Processing
Systems 36 (2023) 53728–53741.
[45] D. Brown, Rank-BM25: A Collection of BM25 Algorithms in Python, 2020. URL: https://doi.org/10.</p>
      <p>5281/zenodo.4520057. doi:10.5281/zenodo.4520057.
[46] T. Dettmers, A. Pagnoni, A. Holtzman, L. Zettlemoyer, Qlora: eficient finetuning of quantized
llms, in: Proceedings of the 37th International Conference on Neural Information Processing
Systems, NIPS ’23, Curran Associates Inc., Red Hook, NY, USA, 2023.
[47] J. Wei, X. Wang, D. Schuurmans, M. Bosma, B. Ichter, F. Xia, E. H. Chi, Q. V. Le, D. Zhou,
Chainof-thought prompting elicits reasoning in large language models, in: Proceedings of the 36th
International Conference on Neural Information Processing Systems, NIPS ’22, Curran Associates
Inc., Red Hook, NY, USA, 2022.
[48] L. Cripwell, J. Legrand, C. Gardent, Simplicity level estimate (sle): A learned reference-less metric
for sentence simplification, in: Proceedings of the 2023 Conference on Empirical Methods in
Natural Language Processing, 2023, pp. 12053–12059.
[49] F. Alva-Manchego, C. Scarton, L. Specia, The (un) suitability of automatic evaluation metrics for
text simplification, Computational Linguistics 47 (2021) 861–889.
[50] L. Cripwell, J. Legrand, C. Gardent, Evaluating document simplification: On the importance of
separately assessing simplicity and meaning preservation, LREC-COLING 2024 (2024) 1.
[51] A. de Wynter, A. Hevia, S.-Q. Chen, A user-centered evaluation of spanish text simplification,
arXiv preprint arXiv:2308.07556 (2023).</p>
    </sec>
    <sec id="sec-11">
      <title>A. Prompts</title>
      <p>Task 1 Initial Adaptation Prompt
Eres un asistente que adapta textos a un estilo de lenguaje claro.</p>
      <p>Usas un lenguaje accesible para todo el público, evitando jerga y tecnicismos.
Usas palabras simples y frases cortas.</p>
      <p>Estructuras el texto de forma lógica, con párrafos cortos y bien separados.</p>
      <p>Utilizas negritas, cursivas y otros recursos tipográficos
para destacar información importante, pero sin abusar de ellos.
[Translation]</p>
      <p>You are an assistant who adapts texts to plain language.</p>
      <p>You use language accessible to all audiences, avoiding jargon and technical terms.
You use simple words and short sentences.</p>
      <p>You structure your text logically, with short, well-spaced paragraphs.</p>
      <p>You use bold, italics, and other typographical features to highlight important information,
but without overusing them.</p>
      <p>Task 2 Initial Adaptation Prompt
Eres un asistente que adapta textos a un estilo de lectura fácil para personas con discapacidad
o adultos con dificultades lectoras. Usas siempre frases cortas de menos de 15 palabras
y palabras simples. Evitas usar oraciones pasivas, relativas o coordinadas.</p>
      <p>Añades explicaciones para las palabras y conceptos más complejos al final del texto,
en una sección titulada "Palabras difíciles".</p>
      <p>Añades puntuación para separar ideas y hacer que el texto sea más fácil de seguir.
Separas el texto adaptado en secciones con su propio título. Evitas usar porcentajes
y cifras complicadas, remplazandolos por formulas como "más de" o "menos de".</p>
      <p>Adaptas cada contenido original que se te indique sin añadir comentarios sobre lo que has hecho.
[Translation]</p>
      <p>You are an assistant who adapts texts to an easy-to-read style for people with disabilities
or adults with reading difficulties. You always use short sentences of fewer than 15 words
and simple words. You avoid using passive, relative, or coordinate clauses.</p>
      <p>You add explanations for more complex words and concepts at the end of the text,
in a section titled "Difficult Words."
You add punctuation to separate ideas and make the text easier to follow.</p>
      <p>You separate the adapted text into sections with their own titles. You avoid using percentages
and complicated figures, replacing them with formulas such as "more than" or "less than".
You adapt each original piece of content as directed without adding comments
about what you have done.</p>
      <p>Task 1 APEC Prompt
Eres un experto en la adaptación de textos a lenguaje claro.</p>
      <p>Aquí están las guías de lenguaje claro:
- Sé breve y conciso.
- Las frases deben de tener la menor cantidad posible de palabras.
- Las palabras deben de tener la menor cantidad posible de sílabas.
- La información breve tiene más efecto. No escribas ni párrafos muy cortos (2 líneas) ni muy largos
(más de 10 líneas). Utiliza el punto seguido para separar ideas.
- Identifica quién hace la acción. Para ello usa la estructura "sujeto + verbo + predicado"
y la voz activa, ya que son más fáciles de entender.
- Conecta las ideas entre sí. Para ello escoge el conector adecuado según su función textual.
- Ordena los elementos.
- Cuando tengas varios elementos, utiliza listas o enumeraciones. Avisa al lector de cuántos elementos
va a encontrar. Por ejemplo: "las cuatro pautas para escribir en lenguaje claro son..."
- Evita palabras técnicas o expresiones complejas.
- En los casos en los que es necesario manteneresos términos, explícalas. Puedes hacerlo
con un paréntesis o con conectores como "es decir".
- Cuidado con abreviaturas y siglas. No des por sentado que las lectoras conocen su significado.
- No elimines ninguna información importante. Ambos estilos son diferentes,
pero el contenido debe ser el mismo.</p>
      <p>Dado este texto de entrada \# Original y adaptación a lenguaje claro # Adaptación, por favor,
indica si la adaptación es correcta o no. Sigue la siguiente estructura:
# Análisis de la adaptación
En este apartado, indica si la adaptación es correcta o no. Si no lo es, explica por qué, indicando
qué aspectos no cumplen con las guidelines de lenguaje claro y qué cambios harías
para corregirlo. Puedes utilizar ejemplos concretos del texto de entrada
y la adaptación para ilustrar tus puntos.
# Corrección
Aquí deberás proporcionar una versión corregida de la adaptación. Si la adaptación es correcta,
simplemente copia la adaptación original. No añadas en esta sección
ningún comentario o explicación adicional. Solo la adaptación corregida.
# Final
En esta sección puedes añadir cualquier comentario o reflexión adicional que consideres relevante.
A continuación te doy el texto de entrada y la adaptación a lenguaje claro a evaluar:
Task 2 APEC Prompt
Eres un experto en la adaptación de textos a lectura fácil
Aquí están las guías de lectura fácil:
- Divide el texto por temas o ideas principales y destaca los puntos importantes.
Puedes dividir o resaltar el texto con viñetas oero no uses demasiadas viñetas juntas.
- Cada frase debe ser lo más breve posible. Más de 15 palabras es más difícil de leer.
- Las palabras deben de tener la menor cantidad posible de sílabas.
- Usa un lenguaje que suene natural cuando se habla.
- Usa frases simples separadas por puntos en lugar de frases complejas con comas
u otros tipos de puntuación.
- Evita usar demasiada puntuación o puntuación compleja como los dos puntos o el punto y coma.
- Asegúrate de que las palabras sean fáciles de entender. Evita la jerga o palabras complicadas
tanto como sea posible.
- Si tienes que usar palabras complicadas (por ejemplo, al explicar un nuevo concepto),
define lo que significan las palabras difíciles con palabras simples.
- Es una buena práctica incluir, al final del texto, una lista de palabras difíciles
si se mantienen en el texto adaptado.
- Es mejor no usar siglas o abreviaturas, pero a veces se pueden usar si son bien conocidas.
- La repetición es mejor que la variedad. Utilice la misma palabra o forma de palabras
cuando se refiera a la misma cosa.</p>
      <p>También puedes introducir frases sobre el mismo tema con la misma forma de palabras.
- Usa donde se pueda aproximaciones como la mitad, un cuarto, 1 de 5, en lugar de porcentajes.
Si usas porcentajes, evita el símbolo % y usa expresiones como "X de cada 100", o "x por ciento".
- Utiliza números enteros, por ejemplo 7% en lugar de 6,8%.</p>
      <p>Dado este texto de entrada # Original y adaptación a lenguaje claro # Adaptación,
por favor, indica si la adaptación es correcta o no. Sigue la siguiente estructura:
# Análisis de la adaptación
En este apartado, indica tu valoración de la adaptación en base a las pautas indicadas:
- Si es correcta, indicalo con: ADAPTACIÓN CORRECTA.
- Si contiene errores críticos, como por ejemplo repeticiones interminables o contenudo sin sentido,
indicalo con: ERRORES CRÍTICOS.
- Si contiene errores que se pueden corregir, indícalo con: ADAPTACIÓN A CORREGIR.
En este caso, indica qué aspectos no cumplen con las pautas de lectura fácil y qué cambios harías
para corregirlo. Puedes utilizar ejemplos concretos del texto de entrada
y la adaptación para ilustrar tus puntos.
# Corrección
Aquí deberás proporcionar una versión corregida de la adaptación en base al análisis que hayas hecho.
Si has identificado que la adaptación es de tipo ERRORES CRÍTICOS, tu corrección deberá ser
una nueva adaptación completa del texto original # Original,
siguiendo las pautas indicadas. Si has identificado que la adaptación
es de tipo ADAPTACIÓN CORRECTA, simplemente copia la adaptación original.</p>
      <p>En cualquier otro caso, proporciona tu versión corregida de la adaptación inicial.
No añadas en esta sección ningún comentario o explicación adicional.
# Final
En esta sección puedes añadir cualquier comentario o reflexión adicional que consideres relevante.
A continuación te doy el texto de entrada y la adaptación a lenguaje claro a evaluar:</p>
    </sec>
    <sec id="sec-12">
      <title>B. APEC Adaptation Examples</title>
      <sec id="sec-12-1">
        <title>La Concejalía de Deportes de Alicante informa que nuestra ciudad acogerá el</title>
        <p>Campeonato Europeo Junior de Natación Artística, durante los días 29 de junio al 3 de julio de 2022.
Se celebrará en las Piscinas Municipales Monte Tossal "José Antonio Chicoy".</p>
        <p>Las entradas se pueden adquirir en la propia Piscina o a través de la página web
de la Real Federación Española de Natación https://rfen.es/es/posts/news/336131</p>
      </sec>
      <sec id="sec-12-2">
        <title>Initial PL adaptation</title>
      </sec>
      <sec id="sec-12-3">
        <title>Alicante acogerá el Campeonato Europeo Junior de Natación Artística</title>
        <p>La Concejalía de Deportes de Alicante informa que la ciudad acogerá el Campeonato Europeo Junior
de Natación Artística del 29 de junio al 3 de julio de 2022.</p>
        <p>El evento se celebrará en las Piscinas Municipales Monte Tossal "José Antonio Chicoy".
Las entradas se pueden comprar en la piscina o en la página web
de la Real Federación Española de Natación.</p>
      </sec>
      <sec id="sec-12-4">
        <title>APEC adaptation</title>
        <p>La ciudad de Alicante organiza un campeonato de natación artística del 29 al 3 de julio de 2022.
El evento se celebra en las piscinas Monte Tossal "José Antonio Chicoy".</p>
        <p>Las entradas se venden allí o en la página web de la
Real Federación Española de Natación: https://rfen.es/es/posts/news/336131.</p>
      </sec>
      <sec id="sec-12-5">
        <title>Human adaptation</title>
      </sec>
      <sec id="sec-12-6">
        <title>La Concejalía de Deportes anuncia que el Campeonato Europeo Junior</title>
        <p>de Natación Artística se celebrará en las Piscinas Municipales Monte Tossal de Alicante
del 29 de junio al 3 de julio del año 2022.</p>
        <p>Las entradas se pueden coger en la piscina o en la página web
de la Real Federación Española de Natación https://rfen.es/es/posts/news/336131
La exposición titulada “Nueva Tabarca, 250 años”, que impulsa la Concejalía de Cultura
del Ayuntamiento de Alicante, en reconocimiento a la aventura humana de los pobladores de la isla,
está conformada por una colección fotográfica, procedente del Archivo Municipal de Alicante,
en colaboración con la Asociación Cultural de Tabarca.</p>
        <p>La muestra, que podrá visitarse hasta el 20 de septiembre, se hace con ocasión
del décimo aniversario y reedición de la revista Canelobre, n° 60
con el título "Tabarca, utopía y realidad", que conllevará también
una mesa redonda en el propio Instituto Gil-Albert, alrededor de Nueva Tabarca,
su pasado, presente y futuro, en colaboración con la propia área de Cultura.</p>
        <p>Este contingente humano terminó su periplo en la isla alicantina de Nueva Tabarca,
después de ser desalojados de su anterior asentamiento en Tabarka,
la isla en aguas de Túnez que fue importante centro de explotación de coral rojo.</p>
        <p>Initial PL adaptation
La exposición “Nueva Tabarca, 250 años” es una muestra que muestra
la historia de los habitantes de la isla de Nueva Tabarca.</p>
        <p>Esta exposición está organizada por la Concejalía de Cultura del Ayuntamiento de Alicante.
La muestra está compuesta por fotos del Archivo Municipal de Alicante
y la Asociación Cultural de Tabarca. Estas fotos muestran la historia de los habitantes de la isla.
La exposición se puede ver hasta el 20 de septiembre. También se publicará
una revista llamada Canelobre. Esta revista tiene el número 60 y el título es “Tabarca, utopía y realidad”.
Además, habrá una mesa redonda en el Instituto Gil-Albert. En esta mesa redonda
hablarán sobre Nueva Tabarca, su pasado, presente y futuro.</p>
        <p>Los habitantes de Nueva Tabarca llegaron a la isla después de ser desalojados
de su antigua casa en Tabarka, una isla en el mar de Túnez.</p>
        <p>En Tabarka, los habitantes explotaban coral rojo.</p>
        <p>APEC adaptation
La exposición "Nueva Tabarca, 250 años" muestra la historia de los habitantes de la isla.</p>
        <p>La Concejalía de Cultura del Ayuntamiento de Alicante organiza esta exposición.</p>
        <p>Las fotos muestran la vida en la isla. Estas fotos vienen del Archivo Municipal de Alicante
y de la Asociación Cultural de Tabarca.</p>
        <p>La exposición está abierta hasta el 20 de septiembre. También se publicará
una revista llamada Canelobre. Es un número especial por su décimo aniversario.</p>
        <p>Esta revista se llama "Tabarca, utopía y realidad".</p>
        <p>Habrá una mesa redonda en el Instituto Gil-Albert sobre Nueva Tabarca.</p>
        <p>Se hablará de su pasado, presente y futuro.</p>
        <p>Los habitantes de la isla llegaron a Nueva Tabarca después de dejar Tabarka.</p>
        <p>Esta isla está en aguas del mar de Túnez. En Tabarka, los habitantes extraían coral rojo.</p>
        <p>Human adaptation
El Instituto Juan Gil Albert presenta la exposición llamada 250 años de Nueva Tabarca
La exposición llamada 250 años de Nueva Tabarca es organizada por la concejalía
de Cultura del Ayuntamiento de Alicante y colabora la Asociación Cultural de Tabarca.</p>
        <p>Esta exposición estará hasta el 20 de septiembre en el Instituto Juan Gil Albert.</p>
        <p>Con esta exposición reconocen las aventuras de las personas que vivieron en la isla de Tabarca.
La exposición tiene fotografías que estaban en el Archivo de Alicante,
que es lugar del ayuntamiento donde guardan documentos de hace muchos años.</p>
        <p>Con esta exposición también celebramos que la revista Canelobre número 60
con el nombre de Tabarca, utopía y realidad, cumple 10 años y hablarán de ella
en Instituto Juan Gil Albert con temas de Nueva Tabarca, su pasado, presente y futuro
en colaboración de la concejalía de Cultura.
El concejal de Educación, Ricardo Recuero, junto a la presidenta de la Orquesta Sinfónica
de Torrevieja (OST), Mamen Mateo, el director de la OST, José Francisco Sánchez, y acompañados
de Jorge Lorente Picó, multinstrumentista, y David Lorenzo, narrador de la obra, han presentado
a segunda edición del Concierto en Familia, que tendrá lugar este próximo domingo, 12 de marzo,
en el Auditorio Internacional de Torrevieja, a las 12:00 horas. Esta segunda edición será un cuento
musicalizado que pretende llevar a todos los asistentes a un viaje inolvidable. A través del narrador,
que encarna el espíritu de la música, contará la historia de un niño que tiene inquietudes musicales
pero que no sabe que instrumento elegir. Picó, el multinstrumentista será el encargado de darle vida
a ese niño que le pedirá deseos al espíritu de la música para saber tocar bien cada instrumento que elija.
La finalidad del Concierto en Familia es acercar a los más pequeños la música clásica y, en especial,
las orquestas sinfónicas. El objetivo de esta iniciativa no es otro que despertar, entre los más jóvenes,
el deseo de aprender a tocar un instrumento. “Para nosotros es un lujo contar con la Orquesta Sinfónica
de Torrevieja para este proyecto. Más aún cuando estamos ante una de las mejores de nuestra
provincia”, han sido las palabras del concejal de Educación, Ricardo Recuero.
“Desde la Orquesta Sinfónica queremos dedicar esté concierto a Tomás Ballester, al estar también
involucrado en este proyecto”, han sido las palabras del maestro José Francisco Sánchez. Las entradas
gratuitas ya están disponibles a través de la web entradasatualcance.com
y en la taquilla del Teatro Municipal.</p>
        <p>Initial ER adaptation
Concierto en Familia en el Auditorio Internacional de Torrevieja
El concejal de Educación, Ricardo Recuero, presentó el Concierto en Familia que tendrá lugar
el 12 de marzo en el Auditorio Internacional de Torrevieja. El concierto es un cuento musicalizado
que llevará a los asistentes a un viaje inolvidable. El multinstrumentista Jorge Lorente Picó dará vida
a un niño que quiere aprender a tocar un instrumento. El objetivo del Concierto en Familia es acercar
a los niños a la música clásica y despertar en ellos el deseo de aprender a tocar un instrumento.
Las entradas son gratuitas y están disponibles en la web y en la taquilla del Teatro Municipal.
APEC adaptation
Concierto en Familia en Torrevieja.</p>
        <p>Ricardo Recuero presenta el Concierto en Familia. Es un cuento para niños pequeños con música y narración.
Lleva a los asistentes a un viaje divertido. Jorge Lorente Picó es un músico que toca varios instrumentos.
David Lorenzo cuenta la historia del concierto. El objetivo es acercar a los niños a la música clásica.
Queremos que aprendan a tocar un instrumento.
¿Cuándo es el concierto? Es el 12 de marzo. Es un domingo y es gratis.
¿A qué hora es el concierto? Es a las doce del día. Las entradas están disponibles en la web y en la taquilla.
Puedes comprarlas en la web o en la taquilla del Teatro Municipal.</p>
        <p>Dónde se pueden comprar las entradas: en la web y en la taquilla.</p>
        <p>Palabras difíciles:
- Orquesta: una banda de música.
- Multinstrumentista: alguien que toca varios instrumentos.
- Narrador: alguien que cuenta la historia.</p>
        <p>Human adaptation
Concierto en Familia Edición 2 en Torrevieja
El concejal de Educación Ricardo Recuero y otras personas importantes presentan el Concierto
en Familia número 2. El concierto es el domingo 12 de marzo en el Auditorio Internacional de Torrevieja.
Es un cuento musical de un niño que le gusta la música pero no sabe que instrumento musical elegir.
El Concierto en Familia quiere que los niños conozcan la música clásica y quieran aprender
a tocar un instrumento. Las entradas son gratis.</p>
        <p>Pide tus entradas en la taquilla del Teatro Municipal o en esta página web: entradasatualcance.com
El concejal de Deportes, Víctor Bernabéu, ha anunciado este miércoles que ante la actual situación
de la alta incidencia en los contagios por COVID-19, y atendiendo a los informes técnicos municipales,
“nos vemos obligados a tener que suspender la edición de JUVE 2022 que teníamos previsto
celebrar los días 2, 3 y 4 de enero en la Glorieta Gabriel Miró”. En este sentido, el edil oriolano
ha manifestado también que los informes técnicos recogen que “no se puede garantizar desde
la administración pública la aplicación de la norma y disminución del riesgo de contagio sin perder
la esencia del evento”. Por otra parte, Bernabéu ha lamentado que “un año más la pandemia nos deja
sin JUVE, pero tenemos que priorizar ante todo la seguridad y la salud de los oriolanos,
esperamos que el próximo año la situación sea diferente y podamos celebrar este evento deportivo
tan esperado por todos”. El año pasado este evento ya no se pudo celebrar precisamente
por la complicada situación de la pandemia y las restricciones establecidas en su momento,
lo que impidió que este evento se pudiera llevar a cabo finalmente.</p>
        <p>Initial ER adaptation
El concejal de Deportes, Víctor Bernabéu, ha anunciado que suspenderá JUVE 2022.
¿Por qué? La razón es la alta incidencia de COVID-19 en la ciudad.
¿Qué dice el concejal? No se puede garantizar la seguridad de los asistentes sin perder la esencia del evento.
¿Qué hace el concejal? Prioriza la seguridad y la salud de los oriolanos.
¿Qué pasó el año pasado? El evento no se pudo celebrar por la pandemia y las restricciones.
Palabras difíciles:
- Incidencia: La incidencia es el número de personas que tienen una enfermedad en un momento determinado.
- Esencia: La esencia es la parte más importante de algo.</p>
        <p>APEC adaptation
El concejal de Deportes suspende JUVE 2022.
¿Por qué? Hay muchos casos de COVID-19 en la ciudad. La situación es complicada y hay gente enferma.
¿Qué dice el concejal? No es seguro sin cambiar las cosas. No se puede garantizar la seguridad sin cambios.
¿Qué hace el concejal? Prioriza la seguridad y la salud. Víctor Bernabéu prioriza la seguridad de todos.
¿Qué pasó el año pasado? El evento no se celebró por la pandemia. La pandemia impidió que JUVE se celebrara.
Palabras difíciles:
- Incidencia: Es el número de enfermos.
- Esencia: Es lo más importante.
La concejalía de Deportes cancela la celebración de la JUVE 2022
El concejal de Deportes anuncia que por los contagios de Covid tienen que suspender la edición de JUVE 2022
que iba a ser el 2, 3 y 4 de enero en la Glorieta Gabriel Miró.</p>
        <p>Es el segundo año que por la pandemia no pueden celebrar la JUVE.</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>S.</given-names>
            <surname>Štajner</surname>
          </string-name>
          ,
          <article-title>Automatic text simplification for social good: Progress and challenges, Findings of the Association for Computational Linguistics: ACL-IJCNLP</article-title>
          <year>2021</year>
          (
          <year>2021</year>
          )
          <fpage>2637</fpage>
          -
          <lpage>2652</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2] https://www.iso.org/standard/78907.html,
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3] https://www.inclusion-europe.eu/easy-to-read/,
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>J.</given-names>
            <surname>Á</surname>
          </string-name>
          .
          <string-name>
            <surname>González-Barba</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Chiruzzo</surname>
            ,
            <given-names>S. M.</given-names>
          </string-name>
          <string-name>
            <surname>Jiménez-Zafra</surname>
          </string-name>
          ,
          <article-title>Overview of IberLEF 2025: Natural Language Processing Challenges for Spanish and other Iberian Languages, in: Proceedings of the Iberian Languages Evaluation Forum (IberLEF 2025), co-located with the 41st Conference of the Spanish Society for Natural Language Processing (SEPLN 2025), CEUR-WS</article-title>
          . org,
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>J.</given-names>
            <surname>Fernández</surname>
          </string-name>
          <string-name>
            <surname>Huerta</surname>
          </string-name>
          , Medidas sencillas de lecturabilidad,
          <source>Consigna</source>
          <volume>214</volume>
          (
          <year>1959</year>
          )
          <fpage>29</fpage>
          -
          <lpage>32</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>B.</given-names>
            <surname>Botella-Gil</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I.</given-names>
            <surname>Espinosa-Zaragoza</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Bonet-Jover</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Madina</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. Molino</given-names>
            <surname>Piñar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Moreda</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I.</given-names>
            <surname>Gonzalez-Dios</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. T.</given-names>
            <surname>Martín</surname>
          </string-name>
          <string-name>
            <surname>Valdivia</surname>
          </string-name>
          , Ureña, Overview of clears at iberlef 2025:
          <article-title>Challenge for plain language and easy-to-read adaptation for spanish texts</article-title>
          ,
          <source>Procesamiento del Lenguaje Natural</source>
          <volume>75</volume>
          (
          <year>2025</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>S.</given-names>
            <surname>Williams</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Reiter</surname>
          </string-name>
          , L. Osman,
          <article-title>Experiments with discourse-level choices and readability</article-title>
          ,
          <source>in: Proceedings of the 9th European Workshop on Natural Language Generation (ENLG-2003) at EACL</source>
          <year>2003</year>
          ,
          <year>2003</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>R.</given-names>
            <surname>Chandrasekar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Doran</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Srinivas</surname>
          </string-name>
          ,
          <article-title>Motivations and methods for text simplification</article-title>
          ,
          <source>in: COLING 1996 Volume 2: The 16th International Conference on Computational Linguistics</source>
          ,
          <year>1996</year>
          . URL: https://aclanthology.org/C96-2183/.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>A.</given-names>
            <surname>Siddharthan</surname>
          </string-name>
          ,
          <article-title>A survey of research on text simplification</article-title>
          ,
          <source>ITL-International Journal of Applied Linguistics</source>
          <volume>165</volume>
          (
          <year>2014</year>
          )
          <fpage>259</fpage>
          -
          <lpage>298</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>S. S.</given-names>
            <surname>Al-Thanyyan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. M.</given-names>
            <surname>Azmi</surname>
          </string-name>
          ,
          <source>Automated text simplification: a survey</source>
          ,
          <source>ACM Computing Surveys (CSUR) 54</source>
          (
          <year>2021</year>
          )
          <fpage>1</fpage>
          -
          <lpage>36</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>G. H.</given-names>
            <surname>Paetzold</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Specia</surname>
          </string-name>
          ,
          <article-title>A survey on lexical simplification</article-title>
          ,
          <source>Journal of Artificial Intelligence Research</source>
          <volume>60</volume>
          (
          <year>2017</year>
          )
          <fpage>549</fpage>
          -
          <lpage>593</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>H.</given-names>
            <surname>Saggion</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Štajner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Ferrés</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. C.</given-names>
            <surname>Sheang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Shardlow</surname>
          </string-name>
          , K. North, M. Zampieri,
          <article-title>Findings of the tsar-2022 shared task on multilingual lexical simplification</article-title>
          ,
          <source>TSAR</source>
          <year>2022</year>
          (
          <year>2022</year>
          )
          <fpage>271</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <surname>K. C. Sheang</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Saggion</surname>
          </string-name>
          ,
          <article-title>Multilingual controllable transformer-based lexical simplification</article-title>
          ,
          <source>Procesamiento del Lenguaje Natural</source>
          <volume>71</volume>
          (
          <year>2023</year>
          )
          <fpage>109</fpage>
          -
          <lpage>123</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>S.</given-names>
            <surname>Narayan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Gardent</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. B.</given-names>
            <surname>Cohen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Shimorina</surname>
          </string-name>
          ,
          <article-title>Split and rephrase</article-title>
          ,
          <source>in: Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing</source>
          ,
          <year>2017</year>
          , pp.
          <fpage>606</fpage>
          -
          <lpage>616</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>C.</given-names>
            <surname>Niklaus</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Cetto</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Freitas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Handschuh</surname>
          </string-name>
          ,
          <article-title>Transforming complex sentences into a semantic hierarchy</article-title>
          ,
          <source>in: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics</source>
          ,
          <year>2019</year>
          , pp.
          <fpage>3415</fpage>
          -
          <lpage>3427</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>D.</given-names>
            <surname>Ponce</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Etchegoyhen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Calleja</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Gete</surname>
          </string-name>
          ,
          <article-title>Split and rephrase with large language models</article-title>
          ,
          <source>in: Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume</source>
          <volume>1</volume>
          :
          <string-name>
            <surname>Long</surname>
            <given-names>Papers)</given-names>
          </string-name>
          ,
          <year>2024</year>
          , pp.
          <fpage>11588</fpage>
          -
          <lpage>11607</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <article-title>Automatic induction of rules for text simplification</article-title>
          ,
          <source>Knowledge-Based Systems</source>
          <volume>10</volume>
          (
          <year>1997</year>
          )
          <fpage>183</fpage>
          -
          <lpage>190</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>F.</given-names>
            <surname>Alva-Manchego</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Bingel</surname>
          </string-name>
          , G. Paetzold,
          <string-name>
            <given-names>C.</given-names>
            <surname>Scarton</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Specia</surname>
          </string-name>
          ,
          <article-title>Learning how to simplify from explicit labeling of complex-simplified text pairs</article-title>
          ,
          <source>in: Proceedings of the Eighth International Joint Conference on Natural Language Processing</source>
          (Volume
          <volume>1</volume>
          :
          <string-name>
            <surname>Long</surname>
            <given-names>Papers)</given-names>
          </string-name>
          ,
          <year>2017</year>
          , pp.
          <fpage>295</fpage>
          -
          <lpage>305</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>