<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Team The Toxinators 2000 at TextDetox CLEF 2025/Multilingual Text Detoxification 2025: The Evolution of Methods for Text Detoxification: The Role of Language in Method Selection</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Andrei Totok</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Artemiy Ermolaev</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Anastasia Izyumova</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Evgeniy Finogeev</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Ecom.tech</institution>
          ,
          <addr-line>Moscow</addr-line>
          ,
          <country country="RU">Russia</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Institute of Precision Mechanics and Optics (ITMO)</institution>
          ,
          <addr-line>Saint Petersburg</addr-line>
          ,
          <country country="RU">Russia</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>National Research Nuclear University (MEPhI)</institution>
          ,
          <addr-line>Moscow</addr-line>
          ,
          <country country="RU">Russia</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <abstract>
        <p>This paper explores various approaches to the task of text detoxication, including the use of lexical resources (such as toxic word dictionaries), methods based on deep learning algorithms (T5, VAE, etc.) and modern large language models (LLMs). The study is conducted on data from PAN: Multilingual Text Detoxication (TextDetox) 2025 - with the aim of identifying the most e‌ective strategies for handling toxic language depending on linguistic specics. It is shown that for low-resource languages, like Tatar, dictionary-based methods show the highest e‌ectiveness. For widely spoken languages, such as English, deep learning methods show better quality. The results highlight the importance of considering linguistic features when selecting a detoxication method and open up possibilities for the further development of adaptive multilingual content ltering systems.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;detoxication</kwd>
        <kwd>pretrained models</kwd>
        <kwd>dictionaries</kwd>
        <kwd>PAN 2025</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>The rapid growth of online communication platforms has led to an increasing amount of user-generated
content, which ofien includes toxic, aggressive, or otherwise harmful language. This phenomenon poses
a serious challenge for maintaining safe and inclusive digital environments. One of the key solutions
to this issue lies in the development of automated systems capable of detecting and neutralizing toxic
language — a task commonly referred to as text detoxication.</p>
      <p>Text detoxication involves transforming a given text in such a way that its toxic or o‌ensive elements
are removed or sofiened, while preserving the original meaning and stylistic coherence. The eld of
NLP has seen rapid development in recent years, enabling diverse approaches to address the problem of
text detoxication.</p>
      <p>
        Each of these approaches comes with its own strengths and limitations. Lexical methods, such as
ltering based on lists of toxic words, o‌er simplicity and speed but ofien fail to account for context,
sarcasm, or subtle forms of toxicity. Deep learning models provide better contextual understanding
when trained on high-quality annotated datasets[
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], yet they may struggle with generalization across
domains and languages[
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. LLMs, particularly those ne-tuned for generation tasks, show good results
in rewriting toxic sentences into non-toxic equivalents while maintaining uency and intent. However,
performance can vary signicantly depending on the linguistic structure and data availability for each
specic language.
      </p>
      <p>
        This paper presents a comparative analysis of di‌erent detoxication strategies across multiple
languages, emphasizing how linguistic characteristics inuence the e‌ectiveness of each method. We
argue that no single approach is universally optimal and that the choice of method should be guided
by the specic properties of the target language. This study utilizes data from PAN: Multilingual
Text Detoxication (TextDetox) 2025[
        <xref ref-type="bibr" rid="ref3 ref4">3, 4</xref>
        ] to explore the most e‌ective strategies for managing toxic
language, taking into account linguistic and cultural di‌erences. We use the Joint metric from PAN,
which consists of three components: Style Transfer Accuracy, Content Preservation, and Fluency. Each
component ranges from 0 to 1, and higher values of both individual components and the nal score
indicate better performance.
      </p>
    </sec>
    <sec id="sec-2">
      <title>2. The formation of toxic words in di‌erent languages</title>
      <p>In modern linguistics, toxic words are commonly referred to as invective vocabulary. The invective
lexicon of a language is enriched through several sources: reinterpretation of existing literary units,
borrowings from foreign languages, and slang. Importantly, invective vocabulary di‌ers from slang in
terms of stability: while slang undergoes frequent changes and updates, the core set of invectives tends
to remain relatively stable over long periods.</p>
      <p>
        As an example, consider American English, where a signicant source of lexical borrowings is
AfricanAmerican vernacular, including such expressions as "motherf*cker" or "of*y". However, invectives should
be distinguished from colloquial speech, which may appear similar but lacks the same emotionally
charged connotation. Colloquial words such as "sh*t" or "a*s", although considered vulgar in formal
contexts, can function as neutral terms in informal speech — especially within certain social groups[
        <xref ref-type="bibr" rid="ref5">5</xref>
        ].
      </p>
      <p>Thus, the meaning and communicative function of the same word depend on the speaker’s social
status and context, leading to an overlap between invective and colloquial vocabulary. This process
facilitates lexical exchange between these two linguistic layers.</p>
      <p>The formation of invective vocabulary can be classied into several key directions. First, many
invectives originate from colloquial, slang, or marginal vocabulary ofien used to express social di‌erences
and conicts. Second, a major contribution to the development of invectives comes from pejoration —
the semantic re-evaluation of literary and conversational words, during which originally neutral or
positive meanings transform into negative ones.</p>
      <p>Linguist I.I. Kremikh identies several types of pejoration, among which metaphorical pejoration is
considered the most productive. It is based on the transfer of negative meaning to an object through
similarity or association. Examples include the Spanish words "globo" ("balloon"), used metaphorically
to refer to a fat person, and "buzo´n de correos" ("mailbox"), used to describe someone with a large
mouth.</p>
      <p>Of particular importance in invective vocabulary are zoomorphisms — metaphors in which people
are compared to animals based on stereotypical associations. For instance, the Spanish words "cerdo"
and "cochino" (both meaning "pig") or the English "pig" and "dirty dog" are used to describe messy or
rude individuals. Similarly, repurposed names of fruits and vegetables, such as "melo´n" ("melon") and
"carrot" (used derogatorily for red-haired people), illustrate how stable phonetic forms can undergo
substantial semantic transformation and shifi in communicative function.</p>
      <p>
        An interesting phenomenon is the positive use of invectives. In some cases, these expressions are
used to convey admiration or praise, reecting their deep integration into everyday speech and the
wide range of emotions they can convey. For instance, in the Tatar language, the word "эт" ("dog"),
which functions as an invective, appears in the phrase "Ах, эт, эшне ничек оста хәл иткэн!" ("Wow,
what a masterful way to solve the problem!"), where it serves as a compliment to one’s skills. Such
expressive usage of invectives is also widely employed in literature to reect natural, vivid speech[
        <xref ref-type="bibr" rid="ref6">6</xref>
        ].
      </p>
      <p>Ethnolinguistic factors play a signicant role in the formation and functioning of invective vocabulary,
especially evident when comparing related or contact languages. The Tatar language, with its rich
historical and cultural layer, contains numerous invectives whose semantics are closely tied to collective
memory and religious identity. A comparative analysis of Tatar invectives with corresponding units in
Bashkir, Chuvash, Udmurt, Mari, and Turkish — a fellow Turkic language — reveals consistent patterns
in the development of o‌ensive vocabulary and its national-cultural characteristics.</p>
      <p>For example, in Tatar, the lexemes "таре" ("christian cross", "damned") and "чукынган" ("christened",
and in an invective sense — "damned") emerged under the inuence of the Christianization of Volga
Tatars in the 16th–18th centuries. Originally neutral or positively connotated, these words transformed
into strong invectives with clearly negative semantics within the Muslim community. The invective
"чукынган" demonstrates high derivational productivity, generating forms such as "чукынчык"
("scoundrel"), "чукынып кит" ("disappear!", "go to hell!"), "чукынды" ("all is lost"), and "чукынгыры"
("may he perish"), which can serve as insults, curses, or exclamatory interjections.</p>
      <p>It is important to note that the strength of the invective tone of these words depends on the religious
and cultural identity of the addressee: for Muslims, they carry a clearly negative nuance, while for
Christians, they retain a neutral or even positive meaning. Similar semantic shifis are observed in
Russian as well (e.g., "нехристь" ("heathen") and "креста на тебе нет" ("no cross, no crown"), originally
used to refer to people of di‌erent faiths).</p>
      <p>Thus, invective vocabulary serves not only as a means of expressing insult but also as a marker of
deep cultural-historical layers and boundaries of linguistic consciousness, reecting complex processes
of identity interaction within society.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Lexical methods</title>
      <p>One of the simplest and fastest ways to reduce the level of toxicity in text is the dictionary-based
approach, which involves replacing potentially toxic words with neutral equivalents or completely
removing o‌ensive terms. The replacement can be either full (replacing the entire word) or partial
(applied when a toxic word appears as part of a longer lexical unit).</p>
      <p>
        We use a Multilingual Toxic Lexicon [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] which consists of toxic words and phrases for each language
from the current task. Three types of processing were considered in this section:
• copy base — the text remains unchanged;
• replace — toxic words or their parts are removed or replaced;
• delete — toxic words are removed without replacement.
      </p>
      <p>Copy base method consisted in the fact that the toxic was not changed in any way. In other words, we
looked at the metrics of unmodied sentences. Replace method involved removing the toxic component
from a word. For example, if the dictionary includes the word "f*ck", but "f*cker" not in the toxic
dictionary, only the sux "er" would remain afier processing.</p>
      <p>Delete method consisted of removing an entire word if it contained any toxic substring. For instance,
if the toxic dictionary included the word "f*ck", but "f*cker" was not included, then full word "f*cker"
would be deleted.</p>
      <p>Results of lexical methods presented in Table 1. The values in the table are given for the Joint metric.</p>
      <p>However, in some languages, such as Ukrainian and Japanese, the replace method demonstrated
signicantly better metrics. This e‌ect may be attributed to specic contextual features that inadvertently
amplify the negative tone of the text during word replacement. Delete method showed signicantly
better results for the Arabic language. Therefore, when applying automated detoxication methods, it
is essential to take into account the linguistic structure and contextual usage of words.</p>
      <p>The highest level of toxicity was observed when using the copy base method, where the input text
remained unchanged. Although this approach preserves the original meaning of the message, it does
not address the task of reducing toxicity and should only be used as a baseline for comparison.</p>
      <p>The purely dictionary-based approach has several limitations, including an inability to account for
context, the need for regular dictionary updates, and the risk of false positives. Hence, a promising
direction for future work appears to be the combination of lexical ltering with text rewriting models,
which can not only eliminate toxic elements but also restore textual structure and preserve semantic
meaning.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Deep learning approaches</title>
      <p>
        A more recent solution in the eld of text detoxication involves approaches based on deep learning
models, such as T5 (Text-To-Text Transfer Transformer) [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] and its variations. These models are capable
not only of removing or replacing toxic words but also of rewriting the original text while preserving
its semantic structure and stylistic features. Thanks to their training on parallel corpora — where each
toxic sentence has a corresponding non-toxic version — these models can generate more natural and
contextually appropriate outputs.
      </p>
      <p>Such models demonstrate strong performance in neutralizing o‌ensive language without signicantly
altering the meaning of the original message. Their e‌ectiveness is especially evident when they are
ne-tuned on high-quality detoxication datasets. A more detailed discussion of these experiments will
follow in the section on large language models (LLMs).</p>
      <p>
        For the current set of experiments, we used the HRQ-VAE [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] model, which has shown promising
results in text paraphrasing tasks. The model is based on the Variational Autoencoder (VAE) [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]
architecture and is designed for sofi style transfer, including reducing the level of toxicity in text.
      </p>
      <p>
        The experiments were conducted on the English language using both the pretrained version of the
model and several versions ne-tuned on the Paradetox [
        <xref ref-type="bibr" rid="ref11 ref12">11, 12</xref>
        ] parallel dataset. We applied the models
trained on three di‌erent datasets: MSCOCO [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ], Paralex [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ], and QQP [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ], evaluating its ability to
reduce toxicity in input sentences. We also explored how di‌erent training strategies a‌ect performance
whether toxic phrases are included or excluded from paraphrase clusters. A paraphrase cluster is a
group containing sentences that are similar in meaning. For example, the sentences "How do I get to
the nearest metro station?", "Can you tell me how to get to the subway?", and "Where is the closest
metro station?" have similar meanings, so they are placed in the same cluster.
      </p>
      <p>Results of di‌erent HRQ-VAE models for English presented in Table 2.The values in the table are
given for the Joint metric.</p>
      <p>Our ndings show that the model’s performance depends heavily on the quality of the training
data and the strategy used to form paraphrase clusters. In some congurations, the model was able to
signicantly reduce toxicity levels, while in others, it failed to fully eliminate harmful content or even
preserved negative connotations.</p>
      <p>In summary, the HRQ-VAE model demonstrates potential for use in text detoxication, particularly
when ne-tuned on carefully curated parallel datasets. However, like other methods, it requires careful
tuning and contextual awareness to avoid unintended preservation or amplication of toxic tone.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Large language models</title>
      <p>A modern stage in the development of text models is represented by large language models (LLMs) such
as ChatGPT [16], Llama3 [17] and others. These models possess signicant capacity for understanding
and generating natural language, making them promising candidates for solving complex tasks, including
text detoxication. Unlike traditional approaches based on rigid rules or simple neural architectures,
LLMs are capable not only of removing or replacing potentially toxic words, but also of rewriting
phrases while preserving the original meaning, style, and logical coherence.</p>
      <p>One of the key advantages of LLMs is their ability to interpret user instructions through prompts,
which enables exible control over the generation process. For example, prompts such as "Make this
more neutral" or "Paraphrase this sentence" allow the model to understand that the tone of the text needs
to be adjusted without sacricing its informational content. This makes LLMs particularly suitable for
use in real-world content ltering systems, where both the reduction of toxicity and the maintenance
of communication quality are essential.</p>
      <p>In the course of this study, several versions of the Flan-T5 [18] model (small, large) were tested.
These models were initially trained on the Paradetox dataset and further ne-tuned on a mixed dataset
that included Paradetox, Parallel Detoxication Dataset small [19], Paranmt-for-detox [20], and
Filtered_paranmt [21]. The training was conducted using di‌erent prompts, such as "Make more neutral"
and "Detoxify", which allowed us to assess the impact of prompt formulation on the quality of text
rewriting. mT0 [22] model from the TextDetox 2024 Multilingual Text Detoxication task.</p>
      <p>Results of di‌erent prompts presented in Table 3. The values in the table are given for the Joint
metric.</p>
      <p>The results demonstrated that the Flan-T5-large variant outperformed smaller versions in terms
of detoxication performance. Specically, afier 4696 training steps, the joint metric reached 0.70497
when using the "Make more neutral" prompt, compared to 0.68628 for Flan-T5-small. This indicates that
increasing the number of model parameters has a positive e‌ect on its ability to reduce textual toxicity.</p>
      <p>It is also worth noting that T5-based models can be adapted to specic languages and domains. For
instance, the spivavtor [23] model, designed for the Ukrainian language, showed improved results when
prompted with "Перефразуйте" ("Paraphrase"). It achieved a joint metric of 0.33832, signicantly lower
than that of dictionary-based approaches. This supports the conclusion that T5-based models have
strong potential for application in multilingual environments.</p>
      <p>Results for spivavtor models on Ukrainian language presented in Table 4. The values in the table are
given for the Joint metric.</p>
      <p>Additionally, we evaluated the Llama3[24] model with varying temperature settings, di‌erent prompts,
and multiple LoRA adapters [25], which allowed us to explore the impact of both generation parameters
and architectural modications on detoxication quality. In our experiments, pre-trained versions of
Llama3 were enhanced with adapters trained on parallel datasets of toxic and non-toxic text.</p>
      <p>Prompts were carefully selected to explicitly instruct the model to reduce toxicity while preserving
the original meaning. Generation was performed using various temperature values (ranging from 0.01
to 0.95), allowing us to evaluate the balance between deterministic and diverse outputs.</p>
      <p>The inuence of temperature on detoxication quality proved to be inversely proportional. Lower
temperature values yielded better results, as the model generated outputs that closely adhered to
the original sentence structure and meaning. In contrast, higher temperature values increased output
variability — although the overall meaning was preserved, the rewritten sentences ofien used completely
di‌erent wording, sometimes reintroducing potentially harmful expressions.</p>
      <p>We also tested several LoRA adapters that di‌ered in size and training data. Interestingly, an adapter
trained on only 50 examples performed slightly better than another trained on 200 examples. However,
the di‌erence between these congurations was minimal, and both still produced outputs that were
relatively close to the base Llama3. Results of llama3-70B model presented in Table 5. The values in the
table are given for the Joint metric.</p>
    </sec>
    <sec id="sec-6">
      <title>6. Leaderboard overview</title>
      <p>Our team, The Toxinators 2000, participated in the Multilingual Text Detoxication task at PAN 2025.
The developed system combined dictionary-based ltering with generative rewriting using a large
language model (LLMs) mT0. According to the evaluation on the platform, we ranked 5th out of 32
teams, achieving an average score of 0.675 across all languages. Additionally, on the LLM-as-Judge
Language
English
Spanish
Deutsch
Chinese
Arabic
Hindi
Ukrainian
Russian
Amharic
Italian
Japanese
Hebrew
French
Tatar
Hindi
leaderboard, we ranked 13th out of 32 teams, achieving an average score of 0.648 across all languages.
The best results were obtained for the following languages on leaderboard:</p>
      <sec id="sec-6-1">
        <title>The full leaderboard is available on the competition page.</title>
      </sec>
    </sec>
    <sec id="sec-7">
      <title>7. Conclusion</title>
      <p>In the course of this study, various approaches to the task of text detoxication were examined, including
the use of lexical resources (dictionaries), deep learning models, and modern large language models
(LLMs). The main objective was to compare the e‌ectiveness of these methods across several languages
— English, Russian, Ukrainian, and others — and to identify optimal strategies for processing toxic
content depending on linguistic characteristics and data availability.</p>
      <p>The mT0 model , when combined with preliminary removal of toxic words and word parts,
demonstrated better performance than the baseline system. For example, in Tatar, the score increased from
0.580 to 0.617 and in Japanese — from 0.582 to 0.644. At the same time, the dictionary-based method
achieved the best results for Hindi, increasing the score from 0.351 to 0.449.</p>
      <p>These ndings conrm that the e‌ectiveness of text detoxication methods is directly inuenced
by linguistic specicity. No single approach proves universally optimal; instead, the choice of method
should be guided by the morphological complexity, cultural context, and data availability of the target
language.</p>
      <p>Thus, future research should focus on developing language-adaptive detoxication systems, which
combine the strengths of dictionary ltering, deep learning, and prompt-based LLM rewriting to ensure
both high-quality output and meaningful reduction of toxic content.</p>
    </sec>
    <sec id="sec-8">
      <title>Declaration on Generative AI</title>
      <sec id="sec-8-1">
        <title>The authors have not employed any Generative AI tools.</title>
        <p>N. Cheng, O. Chernoguz, O. Hart, O. Salpekar, O. Kalinli, P. Kent, P. Parekh, P. Saab, P. Balaji,
P. Rittner, P. Bontrager, P. Roux, P. Dollar, P. Zvyagina, P. Ratanchandani, P. Yuvraj, Q. Liang,
R. Alao, R. Rodriguez, R. Ayub, R. Murthy, R. Nayani, R. Mitra, R. Parthasarathy, R. Li, R. Hogan,
R. Battey, R. Wang, R. Howes, R. Rinott, S. Mehta, S. Siby, S. J. Bondu, S. Datta, S. Chugh, S. Hunt,
S. Dhillon, S. Sidorov, S. Pan, S. Mahajan, S. Verma, S. Yamamoto, S. Ramaswamy, S. Lindsay,
S. Lindsay, S. Feng, S. Lin, S. C. Zha, S. Patil, S. Shankar, S. Zhang, S. Zhang, S. Wang, S. Agarwal,
S. Sajuyigbe, S. Chintala, S. Max, S. Chen, S. Kehoe, S. Sattereld, S. Govindaprasad, S. Gupta,
S. Deng, S. Cho, S. Virk, S. Subramanian, S. Choudhury, S. Goldman, T. Remez, T. Glaser, T. Best,
T. Koehler, T. Robinson, T. Li, T. Zhang, T. Matthews, T. Chou, T. Shaked, V. Vontimitta, V. Ajayi,
V. Montanez, V. Mohan, V. S. Kumar, V. Mangla, V. Ionescu, V. Poenaru, V. T. Mihailescu, V. Ivanov,
W. Li, W. Wang, W. Jiang, W. Bouaziz, W. Constable, X. Tang, X. Wu, X. Wang, X. Wu, X. Gao,
Y. Kleinman, Y. Chen, Y. Hu, Y. Jia, Y. Qi, Y. Li, Y. Zhang, Y. Zhang, Y. Adi, Y. Nam, Yu, Wang,
Y. Zhao, Y. Hao, Y. Qian, Y. Li, Y. He, Z. Rait, Z. DeVito, Z. Rosnbrick, Z. Wen, Z. Yang, Z. Zhao,
Z. Ma, The llama 3 herd of models, 2024. URL: https://arxiv.org/abs/2407.21783. arXiv:2407.21783.
[18] Hugging Face, Flan-t5 models, https://huggingface.co/google/an-t5-small, https://huggingface.</p>
        <p>co/google/an-t5-base, https://huggingface.co/google/an-t5-large, 2025. Accessed: 2025-07-06.
[19] s-nlp, Parallel detoxication dataset (small), https://github.com/s-nlp/parallel_detoxication_data
set/blob/main/parallel_detoxification_dataset_small.tsv, 2025. Accessed: 2025-07-06.
[20] Hugging Face, Paranmt for detox, https://huggingface.co/datasets/s-nlp/paranmt_for_detox, 2025.</p>
        <p>Accessed: 2025-07-06.
[21] s-nlp, Detox releases, https://github.com/s-nlp/detox/releases/, 2025. Accessed: 2025-07-06.
[22] Hugging Face, mt0-xl-detox-orpo, https://huggingf ace.co/s- nlp/mt0- xl- detox- orpo, 2025.</p>
        <p>Accessed: 2025-07-06.
[23] Hugging Face, Spivavtor models, https://huggingf ace.co/grammarly/spivavtor-xxl, https:
//huggingface.co/grammarly/spivavtor-large, 2025. Accessed: 2025-07-06.
[24] Hugging Face, Meta-llama-3-70b, https://huggingf ace.co/meta-llama/Meta-Llama-3-70B, 2025.</p>
        <p>Accessed: 2025-07-06.
[25] E. J. Hu, Y. Shen, P. Wallis, Z. Allen-Zhu, Y. Li, S. Wang, L. Wang, W. Chen, Lora: Low-rank
adaptation of large language models, in: Proceedings of the 10th International Conference on
Learning Representations (ICLR), 2022.</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>A.</given-names>
            <surname>Vaswani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Shazeer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Parmar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Uszkoreit</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Jones</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. N.</given-names>
            <surname>Gomez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Kaiser</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Polosukhin</surname>
          </string-name>
          , Attention is all you need,
          <year>2023</year>
          . URL: https://arxiv.org/abs/1706.03762. arXiv:
          <volume>1706</volume>
          .
          <fpage>03762</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>N.</given-names>
            <surname>Arivazhagan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Bapna</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Firat</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Lepikhin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Johnson</surname>
          </string-name>
          , M. Krikun,
          <string-name>
            <given-names>M. X.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Cao</surname>
          </string-name>
          , G. Foster,
          <string-name>
            <given-names>C.</given-names>
            <surname>Cherry</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Macherey</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <article-title>Massively multilingual neural machine translation in the wild: Findings and challenges, 2019</article-title>
          . URL: https://arxiv.org/abs/
          <year>1907</year>
          .05019. arXiv:
          <year>1907</year>
          .05019.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>J.</given-names>
            <surname>Bevendor‌</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Dementieva</surname>
          </string-name>
          , M. Fro¨be,
          <string-name>
            <given-names>B.</given-names>
            <surname>Gipp</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Greiner-Petter</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Karlgren</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Mayerl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Nakov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Panchenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Potthast</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Shelmanov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Stamatatos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Stein</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Wiegmann</surname>
          </string-name>
          , E. Zangerle, Overview of PAN 2025:
          <article-title>Voight-Kamp‌ Generative AI Detection, Multilingual Text Detoxication, Multi-Author Writing Style Analysis, and Generative Plagiarism Detection</article-title>
          , in: J.
          <string-name>
            <surname>C. de Albornoz</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Gonzalo</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Plaza</surname>
          </string-name>
          ,
          <string-name>
            <surname>A. G. S. de Herrera</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Mothe</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Piroi</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Rosso</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Spina</surname>
          </string-name>
          , G. Faggioli, N. Ferro (Eds.),
          <source>Experimental IR Meets Multilinguality, Multimodality, and Interaction. Proceedings of the Sixteenth International Conference of the CLEF Association (CLEF</source>
          <year>2025</year>
          ), Lecture Notes in Computer Science, Springer, Berlin Heidelberg New York,
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>D.</given-names>
            <surname>Dementieva</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Protasov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Babakov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Rizwan</surname>
          </string-name>
          , I. Alimova,
          <string-name>
            <given-names>C.</given-names>
            <surname>Brune</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Konovalov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Muti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Liebeskind</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Litvak</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Nozza</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. Shah</given-names>
            <surname>Khan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Takeshita</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Vanetik</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. A.</given-names>
            <surname>Ayele</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Schneider</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. M.</given-names>
            <surname>Yimam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Elnagar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mukherjee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Panchenko</surname>
          </string-name>
          ,
          <article-title>Overview of the multilingual text detoxication task at pan 2025</article-title>
          , in: G. Faggioli,
          <string-name>
            <given-names>N.</given-names>
            <surname>Ferro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Rosso</surname>
          </string-name>
          , D. Spina (Eds.),
          <source>Working Notes of CLEF 2025 - Conference and Labs of the Evaluation Forum, CEUR Workshop Proceedings, CEUR-WS.org</source>
          ,
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <surname>Z. N.</surname>
          </string-name>
          ,
          <article-title>Ways of formation of invective vocabulary (based on the material of the piraeus national variant of the spanish language and the american national variant of the english language</article-title>
          ), https://cyberleninka.ru/article/n/puti
          <article-title>-formirovaniya-invektivnoy-leksiki-na-materiale-pireysk ogo-natsionalno-varianta-ispanskogo-yazyka-i-amerikanskogo-</article-title>
          <string-name>
            <surname>natsionalnogo</surname>
          </string-name>
          ,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>D. K.</given-names>
            <surname>Vakhitova</surname>
          </string-name>
          ,
          <article-title>Invective Vocabulary of the Tatar Language: Functional and Ethnomental Aspects</article-title>
          ,
          <source>Ph.d. dissertation</source>
          , Kazan Federal University, Kazan, Russia,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>Hugging</given-names>
            <surname>Face</surname>
          </string-name>
          , Multilingual toxic lexicon, https://huggingface.co/datasets/textdetox/multilingual _toxic_lexicon,
          <year>2025</year>
          . Accessed:
          <fpage>2025</fpage>
          -07-06.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>C.</given-names>
            <surname>Ra</surname>
          </string-name>
          <article-title>‌el</article-title>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Shazeer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Roberts</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Narang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Matena</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. J.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <article-title>Exploring the limits of transfer learning with a unied text-to-text transformer</article-title>
          ,
          <source>Journal of Machine Learning Research</source>
          <volume>21</volume>
          (
          <year>2020</year>
          )
          <fpage>1</fpage>
          -
          <lpage>67</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>T.</given-names>
            <surname>Hosking</surname>
          </string-name>
          , hrq-vae, https://github.com/tomhosking/hrq-vae,
          <year>2025</year>
          . Accessed:
          <fpage>2025</fpage>
          -07-06.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>D. P.</given-names>
            <surname>Kingma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Welling</surname>
          </string-name>
          ,
          <article-title>Auto-encoding variational bayes</article-title>
          , https://arxiv.org/abs/1312.6114,
          <year>2013</year>
          . ArXiv preprint arXiv:
          <volume>1312</volume>
          .
          <fpage>6114</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>V.</given-names>
            <surname>Logacheva</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Dementieva</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Ustyantsev</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Moskovskiy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Dale</surname>
          </string-name>
          , I. Krotova,
          <string-name>
            <given-names>N.</given-names>
            <surname>Semenov</surname>
          </string-name>
          ,
          <string-name>
            <surname>A</surname>
          </string-name>
          . Panchenko, ParaDetox: Detoxi
          <article-title>cation with parallel data, in: Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics</article-title>
          (Volume
          <volume>1</volume>
          :
          <string-name>
            <surname>Long</surname>
            <given-names>Papers)</given-names>
          </string-name>
          ,
          <source>Association for Computational Linguistics</source>
          , Dublin, Ireland,
          <year>2022</year>
          , pp.
          <fpage>6804</fpage>
          -
          <lpage>6818</lpage>
          . URL: https://aclanthology.org /
          <year>2022</year>
          .
          <article-title>acl-long</article-title>
          .
          <volume>469</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>D.</given-names>
            <surname>Dementieva</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Ustyantsev</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Dale</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Kozlova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Semenov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Panchenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Logacheva</surname>
          </string-name>
          ,
          <article-title>Crowdsourcing of parallel corpora: the case of style transfer for detoxication</article-title>
          ,
          <source>in: Proceedings of the 2nd Crowd Science Workshop: Trust</source>
          , Ethics, and
          <article-title>Excellence in Crowdsourced Data Management at Scale co-located with 47th International Conference on Very Large Data Bases (VLDB 2021 (https://vldb</article-title>
          .org/2021/)),
          <source>CEUR Workshop Proceedings</source>
          , Copenhagen, Denmark,
          <year>2021</year>
          , pp.
          <fpage>35</fpage>
          -
          <lpage>49</lpage>
          . URL: http://ceur-ws.
          <source>org/</source>
          Vol-
          <volume>2932</volume>
          /paper2.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <surname>T.-Y. Lin</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Maire</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Belongie</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Bourdev</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Girshick</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Hays</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Perona</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Ramanan</surname>
            ,
            <given-names>C. L.</given-names>
          </string-name>
          <string-name>
            <surname>Zitnick</surname>
          </string-name>
          , P. Dolla´r, Microsofi coco: Common objects in context,
          <year>2015</year>
          . URL: https://arxiv.org/abs/1405.0312. arXiv:
          <volume>1405</volume>
          .
          <fpage>0312</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>A.</given-names>
            <surname>Fader</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Zettlemoyer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Etzioni</surname>
          </string-name>
          ,
          <article-title>Paraphrase-driven learning for open question answering</article-title>
          ,
          <source>in: Proceedings of the 51st Annual Meeting of the Association for Computational Linguistics (ACL)</source>
          ,
          <year>2013</year>
          , pp.
          <fpage>1608</fpage>
          -
          <lpage>1618</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Hamza</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Florian</surname>
          </string-name>
          ,
          <article-title>Bilateral multi-perspective matching for natural language sentences</article-title>
          ,
          <source>in: Proceedings of the 26th International Joint Conference on Articial Intelligence (IJCAI)</source>
          ,
          <year>2017</year>
          , pp.
          <fpage>4144</fpage>
          -
          <lpage>4150</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>