<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>Conference and Labs of the Evaluation Forum, September</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Linguistic_Hygenist at PAN 2024 TextDetox: HybridDetox - A Combination of Supervised and Unsupervised Methods for Efective Multilingual Text Detoxification</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Susmita Gangopadhyay</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>M.Taimoor Khan</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Hajira Jabeen</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>GESIS Leibniz Institute for the Social Sciences</institution>
          ,
          <addr-line>Köln</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2024</year>
      </pub-date>
      <volume>0</volume>
      <fpage>9</fpage>
      <lpage>12</lpage>
      <abstract>
        <p>Text detoxification is the process of revising toxic comments to neutralize their toxicity by eliminating inappropriate content, while preserving the meaning of the message. Toxicity can manifest in various forms, including the use of curse words, insults, hate speech, cyberbullying, or trolling. The present-day social media landscape is rife with toxic comments, necessitating a text detoxification system. Unlike the conventional practice of blocking ofensive content through moderation, detoxification preserves the valuable information contained in the message. This paper details our approach for multilingual text detoxification as part of the Multilingual Text Detoxification (TextDetox) 2024 [1] Challenge organized by the PAN lab [2]. Our approach consists of two components i.e., the Supervised T5-BART Module for English and Russian languages with parallel corpora and the Unsupervised PLM Detoxifier for the other seven languages. The Supervised T5-BART Module uses T5 and BART as base models, with exponentially weighted moving average and ROUGE scores as loss functions for Russian and English, respectively. The Unsupervised PLM Detoxifier utilizes hashing techniques, log odds ratio, and linguistic patterns to identify and conceal toxic words across all languages. Additionally, it incorporates a mask prediction model to maintain the original sentence's meaning intact. Our proposed approach has achieved an average score of 0.315 across all languages, exhibiting outstanding performance in English, German, and Ukrainian for style transfer, content preservation, and fluency.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;PAN 2024</kwd>
        <kwd>text-detoxification</kwd>
        <kwd>toxicity</kwd>
        <kwd>mask-prediction</kwd>
        <kwd>sentence-similarity</kwd>
        <kwd>sequence-to-sequence models</kwd>
        <kwd>CEUR-WS</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Internet access has revolutionized information dissemination, providing unprecedented opportunities
worldwide. However, this rapid and uncontrolled proliferation of information containing user-generated
content could also contain toxic information that is considered harmful, ofensive, or inappropriate.
Text detoxification is a critical endeavor in the contemporary digital landscape, where the proliferation
of toxic comments poses significant challenges to online discourse [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. Detoxification process involves
the meticulous revision of toxic comments to neutralize their toxicity while ensuring that the essence
of the original message remains intact [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. Toxicity can manifest in numerous forms, including the use
of curse words, insults, hate speech, cyberbullying, or trolling, contributing to an unhealthy online
environment [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. This pervasive toxicity underscores the urgent need for efective text detoxification
systems to maintain a healthier online ecosystem [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. Identification of toxicity in text is an active area of
research. Today, social networks such as Facebook, Instagram are trying to address the problem of toxicity.
However, they usually simply block ofensive content through moderation [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. Text detoxification
prioritizes the preservation of valuable information within the message while neutralizing its toxicity.
      </p>
      <p>
        Significant progress has been made in detecting ofensive or toxic speech. The supervised text
detoxification techniques are used for languages with abundant resources having parallel corpora [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ].
On the other hand, the unsupervised techniques target languages with limited resources by employing
alternative methods [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. In the realm of multilingual text detoxification, existing approaches typically
adopt a combination of supervised and unsupervised techniques [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]. The integration of pretrained
models has been instrumental in advancing text detoxification eforts [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]. These models, trained
on vast amounts of data, possess a remarkable ability to understand and generate human-like text
across various languages. However, despite their potential, several open challenges like the inability
to generalize to diferent contexts, ineficiency to handle implicit and subtle toxicity, and below-par
performance in multilingual setup remains [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. Adapting pretrained models to diverse languages and
cultural contexts while ensuring their efectiveness in detecting and neutralizing toxic content presents
a significant hurdle [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] due to the continuously evolving language or presence of sarcasm, innuendo,
or coded language in the text [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ].
      </p>
      <p>Our proposed approach aims to tackle multilingual detoxification by adopting a hybrid method of
supervised named Supervised T5-BART Module and unsupervised modules named Unsupevised PLM
Detoxifier. The Supervised T5-BART Module fine-tunes T5 Seq2Seq model for Russian, which was
originally trained in a teacher-forcing style for multiple NLP tasks like summarization, translation, and
text generation. It uses an exponential weighted moving average (EWMA) score for loss evaluation.
Additionally, it fine-tunes BART model for English using ROUGE scores for loss evaluation. Meanwhile,
the Unsupervised PLM Detoxifier adopts a multi-step process, including the masking of multiple toxic
tokens and predicting a suitable mask replacement while still preserving the meaning. By leveraging
both supervised and unsupervised methods, our approach ofers a robust and versatile solution to the
complex problem of multilingual text detoxification.</p>
      <p>In the subsequent sections, we describe the problem statement, related previous research, our proposed
approach, and present some examples from our results. In addition, we also share our vision of future
work that could be adopted in this research direction.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Problem Definition</title>
      <p>The competition expects a text detoxification system for 9 languages from diferent linguistic families.
Parallel training corpora of several thousand toxic-detoxified pairs are available only for English and
Russian languages. For the remaining 7 languages—Spanish, German, Chinese, Arabic, Hindi, Ukrainian,
and Amharic—only texts containing toxic content were provided. For all 9 languages, a list of prominent
toxic lexicons was provided, varying in number. For languages like English and Russian where parallel
training corpora was available, fine-tuning of any text-generation model was allowed. The main
challenge of this competition was to use a mix of supervised and unsupervised approaches to develop a
multilingual text detoxification system. The evaluation was based on both automatic methods such as
duplication, deletion, and backtranslation as mentioned on the challenge website1 as well as manual
verification of the detoxified text.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Related Work</title>
      <p>There have been numerous studies and shared tasks focusing on toxicity detection, particularly for
English language. One of the earliest and most notable eforts came from several Kaggle competitions
organized by the Jigsaw/Conversation AI team, which included the “Toxic Comment Classification
Challenge”2 in 2018, the “Unintended Bias in Toxicity Classification Challenge” 3 in 2019, and the
“Multilingual Toxic Comment Classification Challenge” 4 in 2021. These competitions provided some
of the largest datasets for English toxicity detection, covering multiple types of toxicity such as toxic,</p>
      <sec id="sec-3-1">
        <title>1https://pan.webis.de/clef24/pan24-web/text-detoxification.html</title>
        <p>2https://www.kaggle.com/c/jigsaw-toxic-comment-classification-challenge
3https://www.kaggle.com/c/jigsaw-unintended-bias-in-toxicity-classification
4https://www.kaggle.com/code/bond005/siamese-xlm-r-for-multilingual-sentiment-analysis
obscene, threat, insult, and identity hate, along with multilingual test sets for other languages like
Spanish, French, Italian, and Russian.</p>
        <p>Starting in 2019, the detection of toxicity and ofensive language has been a major focus at SemEval.
This began with the SemEval-2019 Task 6 and continued with the SemEval-2020 Task 12, both centered
around identifying and categorizing ofensive language in Social Media (OfensEval), which garnered
considerable interest and participation. The emphasis on toxicity persisted with the SemEval-2021 Task
on Toxic Spans Detection. This task was designed to pinpoint the exact spans within a text that make it
toxic, providing valuable assistance to human moderators who have to manage lengthy and potentially
harmful comments.</p>
        <p>In 2022, the arena of toxicity detection continued to buzz with activity, featuring events such as
the “Multimedia Automatic Misogyny Identification (MAMI)” competition. This unique challenge
focused on identifying misogynous memes, utilizing a comprehensive analysis of both textual content
and accompanying images. By shedding light on the pervasive issue of systemic gender inequality
and discrimination against women in online spaces. This competition played a pivotal role in raising
awareness and fostering discussions on these critical issues.</p>
        <p>
          Apart from these shared tasks and competitions, several other research works focus on the task of
toxicity identification and text detoxification. Focus has been on utilizing Deep Learning models like
LSTM [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ], utilization of embedding models [16], and incorporation of context [17] in the detection
of toxic texts. Detoxification is generally framed as a style transfer from toxic to neutral (non-toxic)
style, using parallel datasets labeled for toxicity. For example, Logacheva [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] created such a parallel
corpus for the English language as a resource for utilization in the detoxification task. Researchers such
as Laugier [18] used pretrained text-to-text transformer trained on civil comments dataset to create
lfuent and neutral sentences from toxic ones. Detoxification eforts often rely on style transfer models
tested in other domains. For example, fine-tuning autoencoders with additional style classification and
cycle-consistency losses [18] and applying point-wise corrections and seq2seq models to improve text
lfuency and style [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ].
        </p>
        <p>In terms of multilingual and low-resource languages, significant research has been conducted in
multilingual text generation [19], language agnostic sentence embeddings [20], and translation of
low-resource languages [21]. However, multilingual text detoxification remains a challenge that is
relatively under-explored and is still active. A recent challenge, RUSSE-2022 [22], focused solely on
detoxifying Russian texts. Our approach contributes to this unique and evolving area of research by
proposing a unified pipeline for detoxification across multiple languages, including low-resource ones.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. HybridDetox Pipeline</title>
      <p>We propose a method that allows us to efectively address the challenge of detoxification across all
languages in the dataset. Our proposed methodology is a hybrid of supervised and unsupervised
approaches. Our pipeline takes toxic text as input, processes it, and rephrases it into detoxified text.
Figure 1 demonstrates the entire detoxification pipeline for the languages in study.</p>
      <sec id="sec-4-1">
        <title>4.1. Language Detection Module</title>
        <p>The first step a toxic text passes through is a language detection module. Although sentences and their
corresponding languages were provided in the test data, the language detection module was added to
simulate a real production scenario. We used the Python langdetect5 library for this purpose. If the
detected language is English or Russian the text is forwarded through the Supervised T5-BART Module.
For the remaining seven languages the text is passed through an Unsupervised PLM Detoxifier.</p>
        <p>Input:</p>
        <p>Toxic Text
Detect Language</p>
        <p>EN/RU</p>
        <p>Supervised T5-BART Module</p>
        <p>Output:
Detoxified Text
Other Languages</p>
        <p>Unsupervised PLM Detoxifier</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Supervised T5-BART Module</title>
        <p>Supervised T5-BART Module fine-tunes classifiers for English and Russian having parallel corpora. We
used T5 (Text-to-Text Transfer Transformer) model as our base model for the Russian language [23].
T5 is designed around the innovative concept of having a single architecture across diverse tasks to
benefit from transfer learning. It is trained on large-scale diverse datasets that enhanced its ability to
understand and generate close to human-like text. T5 has demonstrated significant advancement for
multiple NLP text generation-related tasks e.g., translation, text summarization, etc. We fine-tuned T5
on parallel corpora for Russian using exponentially weighted moving average (EWMA) loss function. It
puts more emphasis on the recently generated text and is not efected by extreme values.</p>
        <p>We employed BART for the English parallel corpora and fine-tuned for text detoxification [ 24]. BART
model is efective for diferent text generation and comprehension tasks. Its architecture is flexible that
facilitates fine-tuning for specific tasks with parallel corpora. BART has encoder-decoder architecture
where the encoder is similar to BERT while the decoder is a GPT model. During fine-tuning, the models
are exposed to a labeled dataset containing pairs of toxic and detoxified texts. ROUGE measures is used
as loss measure to evaluate the quality of the generated text. It compares n-grams between generated
texts and label text with higher overlap desired. We computed ROUGE-1, ROUGE-2 and ROUGE-L to
train multiple models where the same measure were used to pick the best model.</p>
      </sec>
      <sec id="sec-4-3">
        <title>4.3. Unsupervised PLM Detoxifier</title>
        <p>Unsupervised PLM Detoxifier supports languages without parallel corpora. It is trained for 7 languages
including low-resource languages i.e., Chinese, Amharic, German, Hindi, Arabic, Ukrainian, and Spanish.
The working of Unsupervised PLM Detoxifier is explained with the following submodules.
4.3.1. Toxic Words Identification and Masking
To identify toxic words in the sentences, we adopted a combination of hashing-based techniques and
log-odds ratio. As a starting point, we utilized the list of toxic lexicons provided in the challenge6. Each
language has a list of toxic lexicons containing prominent curse words specific to that language. We
employed a hashing-based sequence-matching mechanism7 to identify words similar to these lexicons
beyond a certain threshold. These identified toxic words were then removed from the sentences and
replaced with masks. Suitable threshold values 1 are identified based on manual evaluation.</p>
        <p>In the next step, our approach relied on the principle that curse words are relatively rare and would
appear less frequently in a neutral corpus compared to a toxic one. Therefore, the log-odds ratio between
any normal neutral corpus and a toxic corpus would highlight a list of toxic words. The log-odds ratio</p>
        <sec id="sec-4-3-1">
          <title>6https://huggingface.co/datasets/textdetox/multilingual_toxic_lexicon 7https://docs.python.org/3/library/diflib.html</title>
          <p>defines the relative frequency comparison and measures how often or less frequent a word is in one
corpus compared to another. A higher log-odds ratio indicates that the word is much more common in
the target corpus (e.g., toxic text) than in the reference corpus (e.g., neutral text). Mathematically, the
log-odds ratio for a word  can be defined as:
where ( | ) and ( | ) are the occurrence probabilities of the word  in toxic and
neutral corpora, respectively.</p>
          <p>
            In summary, the log-odds ratio helps identify words that are significantly more likely to appear in
toxic texts compared to neutral ones, thus aiding in the detection of toxic language. We utilized the
development set’s toxic and neutral pairs for this experiment, but it could also have been conducted
with any toxic and neutral corpus in the target languages. From the extracted list of words and their
log-odds ratios, we selected those with a score above 2 as toxic words. This threshold was chosen
because the log-odds ratio values are in range [
            <xref ref-type="bibr" rid="ref1">0, 1</xref>
            ]. We aimed to maintain a balanced value to ensure
that we accurately identify toxic words while minimizing false positives and negatives. Additionally,
we cleansed our generated toxic lexicon list by filtering out stopwords and words that are less than
3 characters long. This was done to eliminate special characters, symbols, or incomplete random
words. The filtering criteria for removing unwanted stopwords was based on general observation.
This approach ensured that we efectively masked words likely to be curse words, thereby excluding
stopwords and special characters that might have got on to the list of toxic words.
4.3.2. Mask Placement with Linguistic Patterns
Languages follow certain grammatical paradigms or linguistic rules that aid in constructing sentences.
By observing these rules, we were able to better process the masks in sentences. We found that for any
language if curse words appear at the beginning or end of a sentence, they can be simply removed.
Additionally, when multiple consecutive masked words were present, they could be combined into
a single mask without losing the overall meaning of the sentence. Figure 2 shows an example of our
implemented linguistic paradigms. For ease of understanding, the provided example is in English.
4.3.3. Mask Prediction
Following the process of identifying and masking toxic words, and implementing linguistic rules, we
were left with sentences containing masked toxic words. To handle these, we used the XLM-RoBERTa
large model [25], which is pretrained in a self-supervised manner on 2.5TB of filtered CommonCrawl
data spanning 100 languages, including all languages featured in the competition. The model employs a
Masked Language Modeling (MLM) objective. It randomly masks 15% of the words in the input sentence,
processes the entire masked sentence through the model, and predicts the masked words. We chose this
model because, unlike traditional recurrent neural networks (RNNs) that process words sequentially or
autoregressive models like GPT that internally mask future tokens, XLM-RoBERTa learns a bidirectional
representation of the sentence. Using this model, we predicted the top three probable replacements for
each mask and generated sentences accordingly. For sentences with multiple masks, this resulted in 3
possible sentences.
4.3.4. Sentence Similarity
From our resultant 3 sentences generated from the masked predictions, we used a sentence transformer
model [26] to generate embeddings for each of the sentences along with their parallel toxic input sentence.
The model works in a way that sentences with similar meanings are associated with embeddings that
are close in the vector space. Then, the semantic textual similarity between two sentences is computed,
and we have sentence pairs annotated together with a score indicating the similarity between them. The
model uses a Siamese network structure that was trained using CosineSimilarityLoss [27]. Among all the
sentence pairs generated, we chose the one that had the lowest score indicating the resultant sentence
closest to the input toxic sentence as our selected output sentence. The code for both Supervised8 and
Unsupervised9 pipelines are made available at GitHub.
          </p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Results</title>
      <p>The experimental setup consists of setting up the threshold values. For masking toxic words in the
unsupervised module, the threshold 1 is set to 0.8. While the threshold 2 for identifying toxic words
based on their log-odds ratio is set to 0.5. The threshold values are determined based on manual
evaluation. In supervised learning, English has 19744 and Russian has 11090 training samples. Our
method achieves an average score of 0.315 on the leaderboard’s automatic evaluation securing 22nd
position and comparable average results with the mT5 baseline. Notably, we observe exceptional
performance in languages such as English, German, Ukrainian, and Arabic, with scores of 0.47,0.41,0.42,
and 0.52 respectively. In the manual evaluation conducted via crowdsourcing on a random subsample
of 100 texts per language, our method secured the 18th place with an average score of 0.50.</p>
      <p>Upon observing the results of both automatic and manual evaluation, we found that our proposed
approach demonstrated suboptimal performance in Chinese and Spanish and was particularly inefective
for Russian. Despite the loss function indicating convergence and the text being detoxified, manual
evaluation for Russian revealed that the generated text lacked meaningfulness. Further exploration
of our method revealed that the fine-tuned T5 multi-task text generator model that was used for our
method generated smaller tokens resulting in the generation of out-of-vocabulary words. This raises
a significant concern that although language models trained on multilingual text generation may
exhibit promising scores and reduced loss metrics, verifying their efectiveness for languages outside
the researcher’s linguistic proficiency remains challenging. This issue underscores the necessity of
incorporating native speakers in the evaluation process to ensure the semantic integrity of the detoxified
text.</p>
      <p>We also found that using BART with ROUGE for English performed much better on the test set than
using T5 with EWMA for the Russian language. In general, these findings indicate that unsupervised
approaches to multilingual text detoxification using pretrained language models hold promising results
despite the lack of parallel training corpora. Table 1 shows training and validation loss of Supervised
T5-BART Module. Samples of toxic sentences and their detoxified sentence for all 9 languages involved
in the study are given in Figure 3.</p>
      <sec id="sec-5-1">
        <title>8https://github.com/taimoorkhan-nlp/RuEn-supervised-detoxifier 9https://github.com/susmita3107/mDetoxifier-Multilingual-unsupervised-text-detoxifier</title>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>6. Conclusion and Future Work</title>
      <p>In this work, we propose a novel approach that combines both supervised and unsupervised methods
for text detoxification across nine languages, including some low-resource ones. Our work forms a
part of the CLEF Multilingual TextDetox challenge 2024, achieving an average score of 0.315 on the
leaderboard’s automatic evaluation and 0.50 in manual evaluation. While our results are promising, we
acknowledge that there is room for improvement. In the future, we aim to explore diverse methodologies,
such as leveraging multilingual embedding features to identify linguistic similarities among diferent
languages. Additionally, we intend to experiment with various clustering techniques to investigate
potential hierarchical relationships among toxic words. Furthermore, we plan to explore domain
adaptation and transfer learning methods, particularly for languages that share similar roots e.g., Italian,
Spanish, Portuguese and Latin. We anticipate that models trained on languages with similar linguistic
roots might efectively perform on others with comparable linguistic characteristics.
Intelligent Systems and Computer Vision (ISCV), 2022, pp. 1–7. doi:10.1109/ISCV54655.2022.
9806096.
[16] J. Pennington, R. Socher, C. D. Manning, Glove: Global vectors for word representation, in:
Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP),
2014, pp. 1532–1543.
[17] J. Pavlopoulos, J. Sorensen, L. Dixon, N. Thain, I. Androutsopoulos, Toxicity detection: Does
context really matter?, arXiv preprint arXiv:2006.00998 (2020).
[18] L. Laugier, J. Pavlopoulos, J. Sorensen, L. Dixon, Civil rephrases of toxic texts with self-supervised
transformers, arXiv preprint arXiv:2102.05456 (2021).
[19] Z. Wang, S. Mayhew, D. Roth, et al., Extending multilingual bert to low-resource languages, arXiv
preprint arXiv:2004.13640 (2020).
[20] M. Artetxe, H. Schwenk, Massively multilingual sentence embeddings for zero-shot cross-lingual
transfer and beyond, Transactions of the association for computational linguistics 7 (2019) 597–610.
[21] S. Ranathunga, E.-S. A. Lee, M. Prifti Skenduli, R. Shekhar, M. Alam, R. Kaur, Neural machine
translation for low-resource languages: A survey, ACM Computing Surveys 55 (2023) 1–37.
[22] D. Dementieva, V. Logacheva, I. Nikishina, A. Fenogenova, D. Dale, I. Krotova, N. Semenov,
T. Shavrina, A. Panchenko, Russe-2022: Findings of the first russian detoxification shared task
based on parallel corpora, Cited by 1 (2022) 114–131.
[23] C. Rafel, N. Shazeer, A. Roberts, K. Lee, S. Narang, M. Matena, Y. Zhou, W. Li, P. J. Liu, Exploring
the limits of transfer learning with a unified text-to-text transformer, Journal of machine learning
research 21 (2020) 1–67.
[24] M. Lewis, Y. Liu, N. Goyal, M. Ghazvininejad, A. Mohamed, O. Levy, V. Stoyanov, L. Zettlemoyer,
Bart: Denoising sequence-to-sequence pre-training for natural language generation, translation,
and comprehension, arXiv preprint arXiv:1910.13461 (2019).
[25] A. Conneau, K. Khandelwal, N. Goyal, V. Chaudhary, G. Wenzek, F. Guzmán, E. Grave, M. Ott,
L. Zettlemoyer, V. Stoyanov, Unsupervised cross-lingual representation learning at scale, CoRR
abs/1911.02116 (2019). URL: http://arxiv.org/abs/1911.02116. arXiv:1911.02116.
[26] N. Reimers, I. Gurevych, Sentence-bert: Sentence embeddings using siamese bert-networks, arXiv
preprint arXiv:1908.10084 (2019).
[27] F. Rahutomo, T. Kitasuka, M. Aritsugi, et al., Semantic cosine similarity, in: The 7th international
student conference on advanced science and technology ICAST, volume 4, University of Seoul
South Korea, 2012, p. 1.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>D.</given-names>
            <surname>Dementieva</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Moskovskiy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Babakov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. A.</given-names>
            <surname>Ayele</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Rizwan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Schneider</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. M.</given-names>
            <surname>Yimam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Ustalov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Stakovskii</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Smirnova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Elnagar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mukherjee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Panchenko</surname>
          </string-name>
          ,
          <article-title>Overview of the multilingual text detoxification task at pan 2024</article-title>
          , in: G. Faggioli,
          <string-name>
            <given-names>N.</given-names>
            <surname>Ferro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Galuščáková</surname>
          </string-name>
          , A. G. S. de Herrera (Eds.), Working Notes of CLEF 2024 -
          <article-title>Conference and Labs of the Evaluation Forum, CEUR-WS</article-title>
          .org,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>J.</given-names>
            <surname>Bevendorf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X. B.</given-names>
            <surname>Casals</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Chulvi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Dementieva</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Elnagar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Freitag</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Fröbe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Korenčić</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Mayerl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mukherjee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Panchenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Potthast</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Rangel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Rosso</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Smirnova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Stamatatos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Stein</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Taulé</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Ustalov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Wiegmann</surname>
          </string-name>
          , E. Zangerle,
          <article-title>Overview of PAN 2024: Multi-Author Writing Style Analysis, Multilingual Text Detoxification, Oppositional Thinking Analysis, and Generative AI Authorship Verification, in: Experimental IR Meets Multilinguality, Multimodality, and Interaction</article-title>
          .
          <source>Proceedings of the Fourteenth International Conference of the CLEF Association (CLEF</source>
          <year>2024</year>
          ), Lecture Notes in Computer Science, Springer, Berlin Heidelberg New York,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>T.</given-names>
            <surname>Davidson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Warmsley</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Macy</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Weber</surname>
          </string-name>
          ,
          <article-title>Automated hate speech detection and the problem of ofensive language</article-title>
          ,
          <source>in: Proceedings of the international AAAI conference on web and social media</source>
          , volume
          <volume>11</volume>
          ,
          <year>2017</year>
          , pp.
          <fpage>512</fpage>
          -
          <lpage>515</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>S.</given-names>
            <surname>Poria</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Cambria</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Hazarika</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Vij</surname>
          </string-name>
          ,
          <article-title>A deeper look into sarcastic tweets using deep convolutional neural networks</article-title>
          ,
          <source>arXiv preprint arXiv:1610.08815</source>
          (
          <year>2016</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Robinson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Tepper</surname>
          </string-name>
          ,
          <article-title>Detecting hate speech on twitter using a convolution-gru based deep neural network</article-title>
          ,
          <source>in: The Semantic Web: 15th International Conference, ESWC</source>
          <year>2018</year>
          , Heraklion, Crete, Greece, June 3-7,
          <year>2018</year>
          , Proceedings 15, Springer,
          <year>2018</year>
          , pp.
          <fpage>745</fpage>
          -
          <lpage>760</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>L.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Caines</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Pete</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Hutchings</surname>
          </string-name>
          ,
          <article-title>Automated hate speech detection and span extraction in underground hacking and extremist forums</article-title>
          ,
          <source>Natural Language Engineering</source>
          <volume>29</volume>
          (
          <year>2023</year>
          )
          <fpage>1247</fpage>
          -
          <lpage>1274</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>P.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Guberman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Hemphill</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Culotta</surname>
          </string-name>
          ,
          <article-title>Forecasting the presence and intensity of hostility on instagram using linguistic and social features</article-title>
          ,
          <source>in: Proceedings of the International AAAI Conference on Web and Social Media</source>
          , volume
          <volume>12</volume>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>V.</given-names>
            <surname>Logacheva</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Dementieva</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Ustyantsev</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Moskovskiy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Dale</surname>
          </string-name>
          , I. Krotova,
          <string-name>
            <given-names>N.</given-names>
            <surname>Semenov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Panchenko</surname>
          </string-name>
          ,
          <article-title>Paradetox: Detoxification with parallel data, in: Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics</article-title>
          (Volume
          <volume>1</volume>
          :
          <string-name>
            <surname>Long</surname>
            <given-names>Papers)</given-names>
          </string-name>
          ,
          <year>2022</year>
          , pp.
          <fpage>6804</fpage>
          -
          <lpage>6818</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>D.</given-names>
            <surname>Dale</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Voronov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Dementieva</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Logacheva</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Kozlova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Semenov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Panchenko</surname>
          </string-name>
          ,
          <article-title>Text detoxification using large pre-trained neural models</article-title>
          ,
          <source>arXiv preprint arXiv:2109.08914</source>
          (
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>G.</given-names>
            <surname>Hassan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Rabah</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Madriaza</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Brouillette-Alarie</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Borokhovski</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Pickup</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Varela</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Girard</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Durocher-Corfa</surname>
          </string-name>
          , E. Danis, Protocol:
          <article-title>Hate online and in traditional media: A systematic review of the evidence for associations or impacts on individuals, audiences</article-title>
          , and communities,
          <source>Campbell systematic reviews 18</source>
          (
          <year>2022</year>
          )
          <article-title>e1245</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>J.</given-names>
            <surname>Devlin</surname>
          </string-name>
          , M.-
          <string-name>
            <given-names>W.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Toutanova</surname>
          </string-name>
          , Bert:
          <article-title>Pre-training of deep bidirectional transformers for language understanding</article-title>
          , arXiv preprint arXiv:
          <year>1810</year>
          .
          <volume>04805</volume>
          (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>J. H.</given-names>
            <surname>Park</surname>
          </string-name>
          , P. Fung,
          <article-title>One-step and two-step classification for abusive language detection on twitter</article-title>
          ,
          <source>arXiv preprint arXiv:1706.01206</source>
          (
          <year>2017</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Khan</surname>
          </string-name>
          , W. Ma, S. Vosoughi, Lone pine at semeval
          <article-title>-2021 task 5: fine-grained detection of hate speech using bertoxic</article-title>
          ,
          <source>arXiv preprint arXiv:2104.03506</source>
          (
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>J.</given-names>
            <surname>Risch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Krestel</surname>
          </string-name>
          ,
          <article-title>Toxic comment detection in online discussions, Deep learning-based approaches for sentiment analysis (</article-title>
          <year>2020</year>
          )
          <fpage>85</fpage>
          -
          <lpage>109</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>M.</given-names>
            <surname>Taleb</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Hamza</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Zouitni</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Burmani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Lafkiar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>En-Nahnahi</surname>
          </string-name>
          ,
          <article-title>Detection of toxicity in social media based on natural language processing methods</article-title>
          , in: 2022 International Conference on
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>