<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Overview of the Multilingual Text Detoxification Task at PAN 2024</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Daryna</forename><surname>Dementieva</surname></persName>
							<email>daryna.dementieva@tum.de</email>
							<affiliation key="aff0">
								<orgName type="institution">Technical University of Munich</orgName>
								<address>
									<settlement>Munich</settlement>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Daniil</forename><surname>Moskovskiy</surname></persName>
							<email>daniil.moskovskiy@skoltech.ru</email>
							<affiliation key="aff1">
								<orgName type="institution">Skolkovo Institute of Science and Technology</orgName>
								<address>
									<settlement>Moscow</settlement>
									<country key="RU">Russia</country>
								</address>
							</affiliation>
							<affiliation key="aff2">
								<orgName type="institution">Artificial Intelligence Research Institute</orgName>
								<address>
									<settlement>Moscow</settlement>
									<country key="RU">Russia</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Nikolay</forename><surname>Babakov</surname></persName>
							<email>nikolay.babakov@usc.ese</email>
							<affiliation key="aff3">
								<orgName type="institution">Universidade of Santiago de Compostela</orgName>
								<address>
									<settlement>Santiago de Compostela</settlement>
									<country key="ES">Spain</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Abinew</forename><forename type="middle">Ali</forename><surname>Ayele</surname></persName>
							<email>abinew.ali.ayele@uni-hamburg.de</email>
							<affiliation key="aff4">
								<orgName type="institution">Universität Hamburg</orgName>
								<address>
									<settlement>Hamburg</settlement>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Naquee</forename><surname>Rizwan</surname></persName>
							<email>nrizwan@kgpian.iitkgp.ac.in</email>
							<affiliation key="aff5">
								<orgName type="institution">Indian Institute of Technology</orgName>
								<address>
									<settlement>Kharagpur</settlement>
									<country key="IN">India</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Florian</forename><surname>Schneider</surname></persName>
							<email>florian.schneider-1@uni-hamburg.de</email>
							<affiliation key="aff4">
								<orgName type="institution">Universität Hamburg</orgName>
								<address>
									<settlement>Hamburg</settlement>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Xintong</forename><surname>Wang</surname></persName>
							<email>xintong.wang@uni-hamburg.de</email>
							<affiliation key="aff4">
								<orgName type="institution">Universität Hamburg</orgName>
								<address>
									<settlement>Hamburg</settlement>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Seid</forename><forename type="middle">Muhie</forename><surname>Yimam</surname></persName>
							<email>seid.muhie.yimam@uni-hamburg.de</email>
							<affiliation key="aff4">
								<orgName type="institution">Universität Hamburg</orgName>
								<address>
									<settlement>Hamburg</settlement>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Dmitry</forename><surname>Ustalov</surname></persName>
							<email>dmitry.ustalov@jetbrains.com</email>
							<affiliation key="aff6">
								<orgName type="institution">JetBrains</orgName>
								<address>
									<settlement>Belgrade</settlement>
									<country key="RS">Serbia</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Elisei</forename><surname>Stakovskii</surname></persName>
							<email>eistakovskii@gmail.com</email>
							<affiliation key="aff7">
								<orgName type="institution">Independent Researcher</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Alisa</forename><surname>Smirnova</surname></persName>
							<affiliation key="aff8">
								<orgName type="institution">Toloka AI</orgName>
								<address>
									<settlement>Lucerne</settlement>
									<country key="CH">Switzerland</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Ashraf</forename><surname>Elnagar</surname></persName>
							<email>ashraf@sharjah.ac.ae</email>
							<affiliation key="aff9">
								<orgName type="institution">University of Sharjah</orgName>
								<address>
									<settlement>Sharjah</settlement>
									<country key="AE">UAE</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Animesh</forename><surname>Mukherjee</surname></persName>
							<email>animeshm@gmail.com</email>
							<affiliation key="aff5">
								<orgName type="institution">Indian Institute of Technology</orgName>
								<address>
									<settlement>Kharagpur</settlement>
									<country key="IN">India</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Alexander</forename><surname>Panchenko</surname></persName>
							<email>a.panchenko@skol.tech</email>
							<affiliation key="aff1">
								<orgName type="institution">Skolkovo Institute of Science and Technology</orgName>
								<address>
									<settlement>Moscow</settlement>
									<country key="RU">Russia</country>
								</address>
							</affiliation>
							<affiliation key="aff2">
								<orgName type="institution">Artificial Intelligence Research Institute</orgName>
								<address>
									<settlement>Moscow</settlement>
									<country key="RU">Russia</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Overview of the Multilingual Text Detoxification Task at PAN 2024</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">D43F5311644560C057A29F5868CAA2AB</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T17:59+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>PAN 2024, Multilingual Text Detoxification, Text Style Transfer, Multilingualism . Panchenko) 0000-0003-0929-4140 (D. Dementieva)</term>
					<term>0009-0006-7943-4259 (D. Moskovskiy)</term>
					<term>0000-0002-2568-6702 (N. Babakov)</term>
					<term>0000-0003-4686-5053 (A. A. Ayele)</term>
					<term>0009-0007-1872-6618 (N. Rizwan)</term>
					<term>0000-0003-4141-1415 (F. Schneider)</term>
					<term>0009-0002-8005-2259 (X. Wang)</term>
					<term>0000-0002-8289-388X (S. M. Yimam)</term>
					<term>0000-0002-9979-2188 (D. Ustalov)</term>
					<term>0000-0003-2265-7268 (A. Elnagar)</term>
					<term>0000-0003-4534-0044 (A. Mukherjee)</term>
					<term>0000-0001-6097-6118 (A. Panchenko)</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Despite different countries and social platform regulations, digital abusive speech persists as a significant challenge. One of the way to tackle abusive, or more specifically, toxic language can be automatic text detoxification-a text style transfer task (TST) of changing register of text from toxic to more non-toxic. Thus, in this shared task, we aim to obtain text detoxification models for 9 languages: English, Spanish, German, Chinese, Arabic, Hindi, Ukrainian, Russian, and Amharic. This paper presents the Multilingual Text Detoxification (TextDetox) task, the underlying datasets, the evaluation setups, the submissions from participants, and the results obtained. Warning: This paper contains rude texts that only serve as illustrative examples.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>The issue of managing toxic speech remains a crucial aspect of human communication and digital violence prevention <ref type="bibr" target="#b0">[1]</ref>, including the mitigation of toxic responses generated by Large Language Models (LLMs) <ref type="bibr" target="#b1">[2]</ref>. The typical approach to dealing with abusive speech on social platforms involves message blocking <ref type="bibr" target="#b2">[3]</ref>. To address this, numerous toxic and hate speech detection models have been developed for different languages, i.e. English <ref type="bibr" target="#b3">[4]</ref>, Spanish <ref type="bibr" target="#b4">[5]</ref>, Amharic <ref type="bibr" target="#b5">[6]</ref>, Code-Mixed Hindi <ref type="bibr" target="#b6">[7]</ref>, and many others <ref type="bibr" target="#b7">[8]</ref>. However, the recent research indicates a necessity for more proactive moderation of abusive speech <ref type="bibr" target="#b8">[9]</ref>. One such approach is text detoxification.</p><p>Within the baselines approaches for automatic text detoxification, multiple unsupervised baselines were created based on ideas of Delete-Retrieve-Generate <ref type="bibr" target="#b9">[10]</ref>, latent style spaces disentanglement <ref type="bibr" target="#b10">[11]</ref>, or conditional generation with Masked Language Modeling <ref type="bibr" target="#b11">[12]</ref>. However, the latest state-of-the-art outcomes, particularly in English, were attained when parallel data and fine-tuning with text-to-text generation models were employed <ref type="bibr" target="#b12">[13,</ref><ref type="bibr" target="#b13">14]</ref>. At the same time, the availability of such a corpus can be a challenge for new languages and cross-lingual transfer techniques should be applied <ref type="bibr" target="#b14">[15]</ref>.  In this shared task, we explored both setups-cross-lingual and multilingual one-providing new multilingual parallel text detoxification dataset for 9 languages <ref type="bibr" target="#b15">[16]</ref>. The remainder of this paper is structured as follows. Section 2 gives an overview of the TextDetox shared task description. Section 3 provides the full overview of the new multilingual parallel text detoxification dataset collection per each language. In the following sections, the evaluation setups essentials are described-baselines in Section 4, automatic evaluation setup in Section 5, and human evaluation setup in Section 6. The submissions from participants are described in Section 7. Section 8 provides the details about final results-both automatic (Section 8.1) and human (Section 8.2) evaluation leaderboards. Finally, Section 9 concludes the paper.</p><p>All the resources produced from the task are listed at the shared task page <ref type="foot" target="#foot_0">1</ref> and are also mentioned in the corresponding sections.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Shared Task Description</head><p>Here, we provide the shared task main definitions-how we understand toxicity, text style transfer task, cross-lingual and multilingual setup, and the competition rules.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1.">Task Definition</head><p>Definition of Toxicity While there can be different types of toxic language in conversations <ref type="bibr" target="#b16">[17,</ref><ref type="bibr" target="#b17">18]</ref>, i.e. sarcasm, hate speech, threats, in this work, we include samples with substrings that are commonly referred to as vulgar or profane language <ref type="bibr" target="#b18">[19,</ref><ref type="bibr" target="#b12">13]</ref> while the whole main message can be both neutral and toxic, but not hateful with direct insult of individuals or groups of people.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Text Detoxification as Text Style Transfer</head><p>In this work, we adapt the formal task definition of the text style transfer described in <ref type="bibr" target="#b19">[20,</ref><ref type="bibr" target="#b20">21,</ref><ref type="bibr" target="#b12">13]</ref>:</p><p>Having a set of style 𝑆 and a corpus of texts 𝐷, a text style transfer (TST) model is a function 𝛼 : 𝑆 × 𝑆 × 𝐷 → 𝐷 that, given a source style 𝑠 𝑠𝑟𝑐 , a target style 𝑠 𝑡𝑔 , and an input text 𝑑 𝑠𝑟𝑐 , produces an output text 𝑑 𝑡𝑔 such that:</p><p>• The style of the text changes from the source style 𝑠 𝑠𝑟𝑐 to the target style 𝑠 𝑡𝑔 and is measured by a style classifier: 𝜎(𝑑 𝑠𝑟𝑐 ) ̸ = 𝜎(𝑑 𝑡𝑔 ), 𝜎(𝑑 𝑡𝑔 ) = 𝑠 𝑡𝑔 ;</p><p>• The content of the source text is saved in the target text as much as required for the task and estimated by a content similarity function: 𝛿(𝑑 𝑠𝑟𝑐 , 𝑑 𝑡𝑔 ) ≥ 𝑡 𝛿 ;</p><p>• The fluency of the target text achieves the required level according to the fluency estimator:</p><formula xml:id="formula_0">𝜓(𝑑 𝑡𝑔 ) ≥ 𝑡 𝜓 ,</formula><p>where 𝑡 𝛿 and 𝑡 𝜓 are the threshold values for the content preservation (𝛿) and fluency (𝜓) functions and can be adjusted to the specific task. In our task, the source style 𝑠 𝑠𝑟𝑐 is toxic and the target style 𝑠 𝑡𝑔 is non-toxic.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Cross-lingual Text Detoxification</head><p>As parallel text detoxification corpora might not be available for any language, one of the important tasks is to explore cross-lingual text detoxification knowledge transfer. In this case, we assume that training data is available for the resource-rich language (i.e. English) and the task is to obtain a text detoxification system for a new language.</p><p>Multilingual Text Detoxification If parallel corpora available for multiple language, then both monolingual text detoxification models per language and multilingual model for all languages can be obtained.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.">Competition Rules</head><p>The share task timeline was divided in to two phases-development and test.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Development Phase</head><p>For the first phase, only training parallel data for English and Russian from previous works <ref type="bibr" target="#b12">[13,</ref><ref type="bibr" target="#b21">22]</ref> aiming to provide participants to explore cross-lingual transfer techniques.</p><p>Test Phase During the test phase, parallel text detoxification corpora were available for all target languages. Participants was invited to submit monolingual and multilingual solutions.</p><p>Leaderboards During both phases, the leaderboards based on automatic evaluation were available. We used Codalab platform <ref type="bibr" target="#b22">[23]</ref> <ref type="foot" target="#foot_1">2</ref> (and TIRA <ref type="bibr" target="#b23">[24]</ref> as a backup platform). However, despite having powerful models capable of classifying texts and embedding their meanings, human judgment remains superior for making final decisions in the text detoxification task <ref type="bibr" target="#b24">[25]</ref>. Thus, based on a test part subset, we performed human evaluation of the participants submissions. The final leaderboard was based on the human judgements results.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Multilingual Parallel Text Detoxification Dataset</head><p>For each of our 9 target languages, we prepared parallel text detoxification corpus. We asked experts and native speakers to contribute for corpora collection. Further, we describe the collection details per each language: English (Section 3.1), Russian (Section 3.2), Ukrainian (Section 3.3), Spanish (Section 3.4), German (Section 3.5), Hindi (Section 3.6), Amharic (Section 3.7), Arabic (Section 3.8), Chinese (Section 3.9). All the instructions per language are available online. <ref type="foot" target="#foot_2">3</ref> We also opensource the obtained resources for the public usage. <ref type="foot" target="#foot_3">4</ref>For all the data, we adapt the concept of English ParaDetox <ref type="bibr" target="#b25">[26]</ref> collection pipeline as it was designed to automate the data collection as well as verification with crowdsourcing. The pipeline consists of three tasks: Task 1: Rewrite text in a polite way Annotators need to provide the detoxified paraphrase of the text so it becomes non-toxic and the main content is saved or to skip paraphrasing if the text is not possible to rewrite in non-toxic way;</p><p>Task 2: Do these sentences mean the same? Check if the content is indeed the same between the original toxic text and its potential non-toxic paraphrase;</p><p>Task 3: Is this text offensive? Verification of the provided paraphrase if it is indeed non-toxic.</p><p>In the same manner, each language stakeholder asked the annotators to rewrite the toxic samples verifying the main three criteria: (i) the new paraphrase should be non-toxic; (ii) the content should be saved as much as possible; (iii) the resulted text should be fluent but may contain some minor mistakes (as the majority of the original toxic samples are examples from posts from social networks). The rewriting of the texts and verification of their quality could have been done either via crowdsourcing or via manual annotation. The main goal for each language was to obtain 1000 parallel pairs that were later splitted into dev and test sets.</p><p>Data Preprocessing For all languages, we maintain the length of samples as sentences of around 5-20 tokens. Also, if a text sample is from a social network, we anonymize any mentioning of usernames and links.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.">English</head><p>For English, we reused the data from English ParaDetox dataset <ref type="bibr" target="#b12">[13]</ref> and additionally manually marked up approximately 400 pairs to form a validation dataset of 1000 examples.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.1.">Input Data Preparation</head><p>For EnParaDetox, the original toxic texts were taken from Jigsaw toxicity identification challenge train dataset <ref type="bibr" target="#b26">[27]</ref>. We have considered only texts labeled as toxic and severe toxic.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.2.">Annotation Process</head><p>The training and validation sets of EnParaDetox were acquired through crowdsourcing via Toloka<ref type="foot" target="#foot_4">5</ref> platform with fluent English speakers. Additionally, we employed annotators who are fluent in English and hold a Masters degree in Computer Science to compile additional samples to the test set.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.">Russian</head><p>The same as for English, there were previously available training and validation data from previous work <ref type="bibr" target="#b21">[22,</ref><ref type="bibr" target="#b27">28]</ref>. We reused this data and manually annotated some additional toxic examples taken from various toxicity datasets.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Input Toxicity Data</head><p>The original toxic samples were taken from two binary toxicity classification Kaggle Toxic Comments datasets <ref type="bibr" target="#b28">[29,</ref><ref type="bibr" target="#b29">30]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.1.">Annotation Process</head><p>The training and validation sets of RuParaDetox were acquired through crowdsourcing via Toloka platform with fluent Russian speakers. Additionally, we employed annotators who are native in Russian and hold a Masters degree in Computer Science to compile additional samples to the test set.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3.">Ukrainian</head><p>We used the data presented in MultiParaDetox paper <ref type="bibr" target="#b27">[28]</ref> providing the main details of data collection:</p><p>Input Toxicity Data For the Ukrainian language, there was no existing binary toxicity classification corpus. Therefore, we filtered explicitly toxic samples containing obscene lexicon from the predefined list <ref type="bibr" target="#b30">[31]</ref> within the Ukrainian Tweets Corpus <ref type="bibr" target="#b31">[32]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3.1.">Annotation Process</head><p>We adapt ParaDetox <ref type="bibr" target="#b25">[26]</ref> collection pipeline and verified the data quality via crowdsourcing. We utilized the Toloka platform for crowdsourcing tasks in Ukrainian. The annotators, who were native Ukrainian speakers, underwent an examination before starting the tasks.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.4.">Spanish</head><p>We used the data presented in MultiParaDetox paper <ref type="bibr" target="#b27">[28]</ref> providing the main details of data collection:</p><p>Input Toxicity Data For Spanish, we selected samples for annotation from three datasets: hate speech detection ones <ref type="bibr" target="#b32">[33,</ref><ref type="bibr" target="#b33">34]</ref> as well as filtered by keywords Spanish Tweets corpus <ref type="bibr" target="#b34">[35]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.4.1.">Annotation Process</head><p>We adapt ParaDetox <ref type="bibr" target="#b25">[26]</ref> collection pipeline and verified the data quality via crowdsourcing. We utilized the Toloka platform for crowdsourcing tasks in Ukrainian. The annotators, who were native Spanish speakers, underwent an examination before starting the tasks.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.5.">German</head><p>German ParaDetox was collected with several annotators with manual quality verification:</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.5.1.">Input Data Preparation</head><p>The German language source data in this work is based on three datasets containing toxic, offensive, or hate speech comments on social media about primarily political events in Germany or the US. For the two datasets from the GermEval 2018 <ref type="bibr" target="#b35">[36]</ref> and GermEval 2021 <ref type="bibr" target="#b36">[37]</ref> shared tasks, we used data from both the test and the train split and filtered it as follows. For the GermEval 2018 data, we only used samples labeled with the coarse class "OFFENSE" whereas for the GermEval 2021 data -which contains different labels -we only used samples annotated with the "Sub1_Toxic" class. The third dataset <ref type="bibr" target="#b37">[38]</ref> was filtered so only samples were kept where both expert annotators classified the samples as hate speech.</p><p>The data from the three datasets was merged and deduplicated via exact string matching. Furthermore, we removed all samples that included less than 5 or more than 30 white-space separated tokens.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.5.2.">Annotation Process</head><p>To create the final parallel detoxified German dataset, we hired two native German annotators. Annotator A is a female born in 1994 who holds a Master of Arts degree in Social Sciences, and Annotator B is a male born in 1992 who holds a Master of Science degree in Computer Science. The data was distributed so that each sample was transcribed by only one of the annotators.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.6.">Hindi</head><p>Hindi dataset was collected manually by a native-speaker annotator gaining data from multiple sources:</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.6.1.">Input Data Preparation</head><p>Input Toxicity Data We used the HASOC dataset created at FIRE 2019 <ref type="bibr" target="#b38">[39]</ref> as source for Hindi language. Contents in this dataset are relevant within Indian subcontinent which are collected from various social media platforms prevalent in India. In this dataset, hostile posts are divided into HATE SPEECH, OFFENSIVE and PROFANE. For curation, posts containing OFFENSIVE and PROFANE contents in train and test splits were used. 1455 PROFANE posts (1237 train + 218 test) and 873 OFFENSIVE posts (676 train + 197 test) were chosen to prepare detoxifiable toxic data for our task.</p><p>Input Preprocessing On a total of 2328 samples, we first performed deduplication via exact string matching. Mentions, links and emojis were also removed as part of this step.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.6.2.">Annotation Process</head><p>Annotation Task(s) The posts after input preprocessing were manually verified. Those with less than 5 white-space separated tokens were removed and which had more than 25 white-space separated tokens were re-framed to bring them down to this limit. Toxicity and meaning of the posts were unchanged during this re-framing. These posts were then bifurcated into detoxifiable and non-detoxifiable labels. The manual re-framing and bifurcation were carried by a NLP researcher with working experience on hate/toxic speech.</p><p>Out of 2328 samples, 1007 samples were marked as detoxifiable. From these detoxifiable samples, we carefully sampled 24 data points and detoxified them. These detoxified samples were evaluated by two experts who are native Hindi language speakers to provide precise samples to the annotators for detoxifying the whole dataset. Annotators were guided based on expert prepared samples and were asked to re-write toxic pairs in a non-toxic manner, keeping the meaning of the original post unchanged. Detoxification was carried out by two annotators and we provide their details in the corresponding subsection.</p><p>Annotators One male NLP researcher working in the field of hate/toxic speech and another female student enrolled in Bachelor's Degree and having working knowledge in Machine Learning, were employed to carry out the detoxification of whole dataset. Both annotators are Indian, native Hindi speakers and are well versed with the topicality covered in the dataset.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.7.">Amharic</head><p>We compiled new Amharic ParaDetox datasets with the following annotation details, based on prior studies of hate and offensive language:</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.7.1.">Input Data Preparation</head><p>The Amharic ParaDetox dataset is derived from merging two pre-existing studies conducted on the X/Twitter datasets <ref type="bibr" target="#b5">[6,</ref><ref type="bibr" target="#b39">40]</ref>. The dataset introduced by Ayele et al. <ref type="bibr" target="#b39">[40]</ref> was initially annotated into categories of hate, offensive, normal, and unsure by three native speaker annotators, with the gold labels determined through a majority voting scheme. In contrast, the dataset presented by Ayele et al. <ref type="bibr" target="#b5">[6]</ref> was annotated by two native speakers, with a third adjudicator annotator deciding the gold labels for instances where there was no majority consensus. We extracted a subset of these datasets labeled as offensive to create the new Amharic ParaDetox dataset and subsequently reworked this subset using new annotators to determine if the messages could be detoxified and to present non-toxic versions of each message.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Input Toxicity Data</head><p>The input toxicity data is entirely sourced from the two previous studies, namely Ayele et al. <ref type="bibr" target="#b5">[6]</ref> and Ayele et al. <ref type="bibr" target="#b39">[40]</ref>, and has been adapted to meet the requirements of the ParaDetox task.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.7.2.">Annotation Process</head><p>Annotation Task(s) We customized the Potato-POrtable Text Annotation TOol<ref type="foot" target="#foot_5">6</ref> and utilized it for the annotation of Amharic ParaDetox dataset. Annotators were provided annotation guidelines, took hands-on practical training, completed independent training tasks before the main annotation task.</p><p>We conducted pilot annotation of 125 sample items with three native Amharic speaker annotators and evaluated the annotation quality with experts and annotators together in a group meeting to improve the understandings of annotators for the main task. The main annotation task comprises of 2,995 tweets, each annotated by one annotator. Annotators were asked to classify each tweet in to two broad categories, detoxifiable and non-detoxifiable. For the detoxifiable category, annotators are asked to detoxify and re-write the text. For non-detoxifiable tweets, annotators choose non-detoxifiable and select reason as; it is hate speech, it is normal speech or indeterminate to decide the label.</p><p>Annotators Annotators have previous hate speech annotation experiences and already holds Masters degree in Computer Science. Only two of the annotators were evolved in the main annotation task, where both of them are university lecturers and have basic knowledge of natural language processing tasks. One of the annotators is from Adama Scinece and Technology University with experience of 15 years of teaching Computer Science, who is female. The other annotator is a male, who has been teaching Computer science over 12 years in Kotebe University of Education.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.8.">Arabic</head><p>Arabic ParaDetox was collected with several annotators with manual quality verification:</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.8.1.">Input Data Preparation</head><p>The Arabic ParaDetox dataset was created by combining parts of several existing datasets along with the Arabic-translated version of the Jigsaw dataset <ref type="bibr" target="#b26">[27]</ref>. It includes the Levantine Twitter Dataset for Hate Speech and Abusive Language (L-HSAB) <ref type="bibr" target="#b40">[41]</ref>, which focuses on Levantine dialects, and the Tunisian Hate and Abusive Speech (T-HSAB) dataset <ref type="bibr" target="#b41">[42]</ref>, which targets Tunisian dialects. It also incorporates the OSACT dataset <ref type="bibr" target="#b42">[43]</ref> and the Arabic Levantine Twitter Dataset for Misogynistic Language (LeT-Mi) <ref type="bibr" target="#b43">[44]</ref>, which specifically addresses gender-based abuse. These resources combine to form the Arabic ParaDetox dataset, aimed at aiding the development of toxicity classifiers capable of handling Arabic content across various dialects and contexts.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.8.2.">Annotation Process Annotators</head><p>The detoxification process was conducted by three annotators, each with a PhD. The team includes two males and one female, all of whom have a strong interest in computational linguistics. These native Arabic speakers possess a deep understanding of the subjects encompassed within the dataset. Each text sample was transcribed by two of the annotators to ensure accuracy and consistency in the data.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.9.">Chinese</head><p>We collected new Chinese ParaDetox datasets with the following annotation details:</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.9.1.">Input Data Preparation</head><p>Input Toxicity Data The Chinese ParaDetox dataset is derived from TOXICN <ref type="bibr" target="#b44">[45]</ref>, a recently released Chinese toxic language dataset. TOXICN was compiled from social media platforms and comprises 12,011 comments addressing several sensitive topics, including gender, race, region, and LGBTQ issues. From this dataset, we extracted a subset based on multiple criteria: the number of toxic words, the ratio of toxic words in the comments, the length of comments, and the toxic scores of comments.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Input Preprocessing</head><p>We set thresholds for the criteria mentioned above: the number of toxic words ranged from 1 to 5, the ratio of toxic words in comments was less than 0.5, and the length of comments ranged from 3 to 50 words, ensuring suitability for annotators to rewrite them. Following these criteria, we extracted 1,516 samples from the training set and 231 samples from the test set.</p><p>Subsequently, we employed a pre-trained toxic classifier <ref type="bibr" target="#b44">[45]</ref> to compute the toxic scores of the selected comments, using a threshold score of 0.978 to filter the candidates. Ultimately, we collected 1,149 samples from the training set and 231 samples from the test set, resulting in a total of 1,380 samples deemed suitable for annotation.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.9.2.">Annotation Process</head><p>Annotation Tasks For data annotation and verification, we employed a specifically designed threetask pipeline: Task 1: Determine if the sentences are toxic or neutral. Annotators were required to choose one of three options: the given sentence is neutral, toxic but can be rewritten, or toxic and cannot be rewritten. The last option was included based on the observation that some toxic texts are impossible to rewrite in a non-toxic manner.</p><p>Task 2: Rewrite sentences in a non-toxic style. Annotators were instructed to create detoxified versions of the toxic sentences identified in Task 1. They were advised to retain the main content of the original sentences and rewrite the toxic words in a polite manner.</p><p>Task 3: Cross-check verification. The rewritten sentences from Task 2 were cross-distributed to different annotators for verification. The goal was to ensure the rewritten sentences were nontoxic and adhered to our guidelines. If annotators selected the "No" option, indicating the sentence did not meet the criteria, a further meta-rewrite process was initiated.</p><p>From the 1,380 toxic samples, 1,031 samples were successfully detoxified and verified, with 861 from the training set and 170 from the test set. The remaining 349 samples were either considered non-toxic or toxic but could not be rewritten.</p><p>Annotators For the detoxification process, we hired three native Chinese annotators. Two female annotators, both aged 22, hold Bachelor's degrees in Engineering, and a male annotator, aged 32, holds a Master's degree in Computer Science. All annotators are native Chinese speakers residing in mainland China, ensuring they deeply understand the Chinese language and the detoxification task.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.10.">Final Dataset Table 1</head><p>The statistics of all ParaDetox datasets used in the shared task. The human detoxified references were collected either via crowdsourcing or locally hired native speaker. For English and Russian, the previously collected train data was available during all shared task's phases. For other languages, 1 000 samples per language were divided correspondingly into development and test parts.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Language</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Source of Toxic Samples</head><p>Annotation The full picture of the collected ParaDetox data for all target languages is presented in Table <ref type="table">1</ref>. While the methods of collecting human annotations vary across languages-some data were gathered via crowdsourcing, others by hiring local native speakers-the quality of the texts was uniformly verified by experts to ensure three key attributes as introduced in <ref type="bibr" target="#b45">[46,</ref><ref type="bibr" target="#b12">13]</ref>: (i) the style of new paraphrases is genuinely non-toxic, (ii) the main content is preserved, and (iii) the new texts are fluent.</p><p>For each language for the shared task's phases:</p><p>• During the development phase: 400 only toxic parts were available for participants to perform cross-lingual experiments.</p><p>• During the test phase: (i) 400 ParaDetox instances were fully released; (ii) participants should provide their final solutions for 600 toxic parts of the test dataset.</p><p>For English and Russian during all phases, additional training parallel datasets were available released from previous work <ref type="bibr" target="#b12">[13,</ref><ref type="bibr" target="#b21">22,</ref><ref type="bibr" target="#b27">28]</ref>. You can find online fully released development part of the data <ref type="foot" target="#foot_6">7</ref> and the test part only toxic instances. <ref type="foot" target="#foot_7">8</ref></p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Baselines</head><p>We provide four baselines for our shared task: (i) trivial Duplicate baseline; (ii) a rule-based Delete approach; (iii) Backtranslation pipeline that reduces the task to the monolingual one; (iv) finally, finetuned for the downstream task on the dev dataset mT5 instance. The code for all the baselines is available online.<ref type="foot" target="#foot_8">9</ref> </p><p>Duplicate Trivial baseline: the output sentence is a copy-paste of the input sentence. This baseline has 1.0 (or 100%) SIM score by definition.</p><p>Delete For the first unsupervised baseline, we perform an elimination of obscene and toxic substrings from a text according to the predefined lists of keywords. For the shared task, we collected and compiled together the lists of such toxic keywords for all target languages based on openly available sources (see Table <ref type="table">2</ref>). The amount of toxic keywords per language differs which displays the diversity of morphological forms and variations of toxicity expressions across languages. For participants and further public usage, we release our compiled list online. <ref type="foot" target="#foot_9">10</ref></p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 2</head><p>The list of the original sources and the corresponding amount of obscene keywords used to compile multilingual toxic lexicon list for our Delete baseline. Backtranslation As for a more sophisticated unsupervised baseline, we perform translation of non-English texts in English with NLLB <ref type="bibr" target="#b18">[19]</ref> instance <ref type="foot" target="#foot_10">11</ref> and then perform detoxification with the fine-tuned on English ParaDetox train part BART <ref type="bibr" target="#b12">[13]</ref> instance. <ref type="foot" target="#foot_11">12</ref>Fine-tuned mT5 Specifically for the test phase, we fine-tuned the multilingual text-to-text generation model mT5 <ref type="bibr" target="#b50">[51]</ref>. We tuned the mT5-XL <ref type="foot" target="#foot_12">13</ref> on the released for the test phase parallel development part of the presented multilingual data.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Language Original</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Automatic Evaluation Setup</head><p>We adopt the monolingual evaluation pipelines from <ref type="bibr" target="#b12">[13,</ref><ref type="bibr" target="#b21">22]</ref> to our multilingual setup and provide the detailed description below. We evaluate the outputs based on three parameters-style of text, content preservation, and conformity to human references-combining them into the final Joint score. The evaluation script is available online. <ref type="foot" target="#foot_13">14</ref>Style Transfer Accuracy (STA) ensures that the generated text is indeed more non-toxic. To prepare a model for this metric that covers our target languages, we subsampled 5 000 samples-2 500 toxic and 2 500 neutral-from toxicity classification corpora for each language (see references in Table <ref type="table">1</ref>) that were not used for ParaDetox data collection. We released <ref type="foot" target="#foot_14">15</ref> this compiled corpus for participants as an additional dataset for experiments and fine-tuned XLM-R <ref type="bibr" target="#b51">[52]</ref> large instance for the binary toxicity classification task. The model is also available for the public usage <ref type="foot" target="#foot_15">16</ref> and is used in the shared task to estimate the level of non-toxicity in the texts.</p><p>Content Similarity (SIM) is the cosine similarity between LaBSE<ref type="foot" target="#foot_16">17</ref> embeddings <ref type="bibr" target="#b52">[53]</ref> of the source texts and the generated texts.</p><p>Fluency (ChrF1) is used to estimate the proximity of the detoxified texts to human references. While in several previous work language acceptability classifiers based on CoLa-like corpora were utilized for fluency estimation <ref type="bibr" target="#b12">[13,</ref><ref type="bibr" target="#b14">15]</ref>, the recent work <ref type="bibr" target="#b24">[25]</ref> also showed that reference-based metrics achieved high correlations with human evaluation. Thus, we use an implementation of ChrF1 score from sacrebleu library <ref type="bibr" target="#b53">[54]</ref>.</p><p>Joint score (J) is the aggregation of the three above metrics. The metrics STA, SIM and ChrF1 are subsequently combined into the final J score used for the final ranking of approaches. Given an input toxic text 𝑥 𝑖 and its output detoxified version 𝑦 𝑖 , for a test set of 𝑛 samples:</p><formula xml:id="formula_1">J = 1 𝑛 𝑛 ∑︀ 𝑖=1 STA(𝑦 𝑖 ) • SIM(𝑥 𝑖 , 𝑦 𝑖 ) • ChrF1(𝑥 𝑖 , 𝑦 𝑖 ),</formula><p>where STA(𝑦 𝑖 ), SIM(𝑥 𝑖 , 𝑦 𝑖 ), ChrF1(𝑥 𝑖 , 𝑦 𝑖 ) ∈ [0, 1] for each text detoxification output 𝑦 𝑖 .</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Human Evaluation Setup</head><p>For the test set, we performed the human evaluation to obtain final judgements on the participants' systems. The details and instructions of the annotation setups are available for the public usage.<ref type="foot" target="#foot_17">18</ref> </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.1.">General setup</head><p>We used Toloka <ref type="foot" target="#foot_18">19</ref> crowdsourcing platform for manual evaluation of automatic detoxification. For each project, the annotators who indicated the knowledge of the necessary language could access the tasks.</p><p>Quality Control First, the annotators underwent a rigorous selection process, starting with a training pool of 20 tasks with predefined answers and instructional feedback. Following this, annotators faced an examination pool comprising 10 tasks with known answers. Only those achieving a minimum accuracy of 65% in both the training and examination pools were permitted to proceed to the main evaluation tasks.</p><p>During the execution of the real tasks, annotators' performance was meticulously monitored using several measures. Response speed was a critical factor; annotators who responded rapidly (less than 15 seconds per page) to three consecutive task pages were permanently banned. Additionally, we evaluated the answers of the annotators to mixed tasks with known true answers. Those who demonstrated low performance, were defined as achieving less than 65% accuracy, were required to undergo re-examination. Furthermore, annotators who skipped three task pages were also banned from the study.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Evaluation Dataset</head><p>We selected 100 random original toxic samples per each language from the test part of our dataset. Each pair was annotated within a separate project dedicated to the respective language. These projects were categorized into three distinct types, which will be described in detail in the following section.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.2.">Annotation projects and corresponding metrics</head><p>In general, the concept of the human evaluation mirrored the approach used in the automatic evaluation. Each project type focused on assessing one of the three key qualities of detoxification; style transfer accuracy, content similarity, or fluency. Below, we provide a detailed description of each project type and its specific focus within the evaluation process.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Style Transfer Accuracy</head><p>To measure style transfer accuracy, we employed a pairwise comparison between the original toxic text and the generated detoxified text. Annotators were tasked with determining which text was more toxic: the left text, the right text, or neither. An illustration of this task can be found in Figure <ref type="figure" target="#fig_5">2</ref>.</p><p>To enhance realism, we randomized the sequence of original and detoxified texts. Annotators' votes were then converted into numerical values using the following logic: if the original toxic text was deemed more toxic, the vote was recorded as 1; if the detoxified text was considered more toxic, the vote was recorded as 0; if "none of the texts" was selected, the vote was recorded as 0.5. Try out tasks as a Toloker and review results</p><p>Look for possible problems that Tolokers might run into and fix them for best results. Check the instructions and the popular screen size. 83% of users use mobile devices when completing tasks</p><p>Which of the texts is more offensive?</p><formula xml:id="formula_2">1 2 3</formula><p>Which of the texts is more offensive? getting kicked by gre practice tests made sure that i never make this same mistake again .</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>1)</head><p>getting my ass kicked by gre practice tests made sure that i never make this same mistake again .  Content Similarity For the content similarity metric, we presented pairs of texts in a fixed order (the toxic phrase followed by the detoxified phrase) and asked annotators to indicate whether the sense of the two texts was similar. Annotators' responses were binary, with "yes" indicating that the sense was similar and "no" indicating otherwise. The responses were mapped to numerical values, with a "yes" response being recorded as 1 and a "no" response being recorded as 0. An illustration of this task can be found in Figure <ref type="figure" target="#fig_7">3</ref>.</p><p>Fluency For the fluency metric, sentences were evaluated individually rather than in pairs. Annotators were asked whether each text was intelligible and correct, with three possible responses: yes, partially, and no. Each response was mapped to a numeric value: "yes" was scored as 1, "partially" as 0.5, and "no" as 0. The fluency score for a pair of texts was determined by comparing the scores of the detoxified text to the original text. If the detoxified text received a higher or equal fluency score, the pair was scored as 1; otherwise, it was scored as 0. An illustration of this task can be found in Figure <ref type="figure" target="#fig_8">4</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Joint Score</head><p>The final scores for each pair within each project were calculated by weighting them according to the skill of the crowdsourcing annotators. This skill was dynamically assigned based on the quality of their responses to both the tasks with known answers and the real tasks. The logic for calculating the Joint score in the manual evaluation mirrors that of the automatic evaluation. still defending their lying and selfish leader.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Yes No</head><p>He deserves every beating and rape he will be subjected to.</p><p>Skip Do these sentences mean the same?  language, the Joint score was determined by multiplying the three individual scores (style transfer accuracy, content similarity, and fluency) using the same formula as in the automatic evaluation.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="7.">Participants</head><p>We received 20 submissions for the development phase leaderboard and 31 submissions for the test phase leaderboard; the final manually evaluated leaderboard was based on 17 submissions who confirmed their participation in the competition. Here, we briefly describe the solutions of our final participants. Each team is presented with its respective leaderboard name (in some cases, additionally, by a username of the corresponding team member that did a submission) and approach used in brackets:</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Team cake, Submission d1n910 (few-shot Kimi.AI) [55]</head><p>The participants achieved the resulting score with a few-short LLM inference by using a two stage process: first, 400 samples from EN and RU provided datasets were used to be detoxified by a proprietary LLM-Kimi.AI <ref type="bibr" target="#b55">[56]</ref> which is a large language model chatbot developed by Moonshot AI, a Beijing-based startup. In the second step, the participants employed newly detoxified samples to construct a prompt where they were included as examples of the desired behavior and the model Kimi.AI, thus, was prompted to perform detoxification in target languages.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Team SINAI, Submission estrella (Tree-of-Thought with GPT-3.5) [57]</head><p>To get the results, Team SINAI employed the Tree-of-Thought prompting strategy based on the OpenAI's model GPT-3.5 <ref type="bibr" target="#b57">[58]</ref>. Given a toxic sentence, the model was prompted to output three options of potential detoxified sentences.</p><p>Then the model was asked to decide in terms of offensiveness, content, and fluency which one out these sentences was detoxified the most appropriate way.</p><p>Team MarSanAI, Submission maryam.najafi (Mistral-7b with PPO) <ref type="bibr" target="#b58">[59]</ref> This team offered a solution only for two languages: English and Russian. A reinforcement learning method was utilized to fine-tune an LLM-Mistral-7b <ref type="bibr" target="#b59">[60]</ref>-coupled with a Proximal Policy Optimization (PPO) <ref type="bibr" target="#b60">[61]</ref> using the implementation from HuggingFace TRL <ref type="bibr" target="#b61">[62]</ref>; the reward was obtained using the provided toxicity classifier.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Team Linguistic_Hygienist, Submission gangopsa (T5 &amp; BART) [63]</head><p>The solution consisted of two components: i) the supervised solution for the English and Russian languages; ii) the unsupervised solution for the other 7 languages. The supervised solution used T5 <ref type="bibr" target="#b63">[64]</ref> and BART <ref type="bibr" target="#b64">[65]</ref> as base models; the exponentially weighted moving average and ROUGE scores were used as loss functions for Russian and English, respectively. The unsupervised solution utilized hashing techniques, log odds ratio, and grammatical rules to identify and conceal toxic words across other 7 languages; additionally, it incorporated a mask prediction model to maintain the original sentences meaning intact.</p><p>Team VitalyProtasov (mT0-large) <ref type="bibr" target="#b65">[66]</ref> In the proposed solution, the team used a text-to-text model-mT0-large <ref type="bibr" target="#b66">[67]</ref>-which was trained on different combinations of languages. In addition, before training, certain filtering techniques were applied to the data.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Team nikita.sushko (mT0-XL) [68]</head><p>The participant used the text-to-text mT0-XL <ref type="bibr" target="#b66">[67]</ref> model that was fine-tuned in two stages. In the first stage, a model was fine-tuned on the parallel data of all languages; this model was used to generate synthetic parallel data from non-parallel samples. The resulting data was cleaned and filtered using a cosine distance between LaBSE embeddings and the toxicity scores by the provided classification models followed by a modification with improved delete approach. At the end, the synthetic and filtered "golden" data were merged into new training set to fine-tune a new instance of the text-to-text multilingual model.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Team SmurfCat, Submission adugeen (mT0-XL) [69]</head><p>Multilingual model mT0-XL <ref type="bibr" target="#b66">[67]</ref> was as well used by this team. First, the model was fine-tuned for text generation using a combination of parallel and translated datasets. The model was further aligned with the Odds Ratio Preference Optimization (ORPO) <ref type="bibr" target="#b69">[70]</ref>. During the inference stage, the best candidate generated by the model was chosen by calculating scores from STA and SIM models.</p><p>Team gleb.shnshn (zero-shot LLaMa-3) This solution was based on a modern open-source LLM-LlaMa3-70B <ref type="bibr" target="#b70">[71]</ref>. The model was prompted using the zero-shot prompting method for the detoxification task.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Team memu_pro_kotow, Submission SomethingAwful (few-shot LLaMa-3 &amp; mT0-XL) [72]</head><p>In this solution, "uncensored" LLaMa3 <ref type="bibr" target="#b70">[71]</ref> was introduced and initialized for every target language except Amharic. Using the recent alignment jailbreaking method by identifying "refusal" directions and subtracting them from model weights <ref type="bibr" target="#b72">[73]</ref>, they used LLaMa3-70B to get predictions using a few-shot prompting strategy. So, the model received 10 examples of detoxification via starting prompt. For the Amharic language, the text-to-text mT0-XL <ref type="bibr" target="#b66">[67]</ref> model was used: the model was fine-tuned on the Amharic parallel dataset.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Team Magnifying_Glass, Submission ZhongyuLuo (Translation &amp; BART-detox, ruT5-detox &amp; Postprocessing) [74]</head><p>The team used a combination of different methods and models depending on the language. For the majority of languages, the participant used a text-to-text encoder-decoder NLLB translation model <ref type="bibr" target="#b18">[19]</ref> to translate data from various languages into English. Then, the translated data was detoxified using the English BART-detox model <ref type="bibr" target="#b12">[13]</ref>. After that, the resulting parallel synthetic data was translated back into the original languages. For Russian, the specifically Russian text-totext model-ruT5-base-detox <ref type="bibr" target="#b21">[22]</ref>-was employed for the detoxification. In the case of Chinese, the participants, firstly, applied filtering of the training dataset, fine-tuned the pre-trained ruGPT3 <ref type="bibr" target="#b74">[75]</ref> model, and applied the Delete method.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Team nlp_enjoyers, Submission shredder67 (mT5) [76]</head><p>The participant employed a text-to-text model mT5 <ref type="bibr" target="#b50">[51]</ref>. The provided multilingual parallel data from the development phase was used for fine-tuning.</p><p>Team NaiveNeuron, Submission erehulka (few-shot LLaMa-3) <ref type="bibr" target="#b76">[77]</ref> The team used a text-to-text Llama3 <ref type="bibr" target="#b70">[71]</ref> which was prompted using a few-shot method.</p><p>Team team0, Submission dkenco (few-shot Cotype-7b) In this case, the team put a stress solely on the English and Russian languages. Two language-specific approaches were used based on Cotype-7b model <ref type="bibr" target="#b77">[78]</ref>. For English, there was employed a zero-shot prompting technique where the prompt included brief instructions for the text detoxification task. For the Russian language, the team used a few-shot approach: the system prompt included brief instructions for the task to be performed as well as five randomly picked samples from the parallel development set. During inference, for both languages, there were applied regular expressions intended as filters.</p><p>Team NLPunks, Submission bmmikheev (few-shot LlaMa-3) This team used a text-to-text Llama3-70B <ref type="bibr" target="#b70">[71]</ref> by with a few-shot prompting method. For English and Russian, the generated output was evaluated manually. For other languages, GPT-3.5 <ref type="bibr" target="#b57">[58]</ref> was used to evaluate outputs. For all languages, the system prompt was formulated in English.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Team Iron Autobots, Submission razvor (few-shot LlaMa-3)</head><p>The participant as well used a text-to-text Llama3-70b <ref type="bibr" target="#b70">[71]</ref> with a few-shot prompting method.</p><p>Team MBZUAI-UnbabelDetox, Submission mkrisnai (few-shot GPT-3.5) In this team, a twostep prompting approach was utilized. At the first step, GPT-3.5 <ref type="bibr" target="#b57">[58]</ref> was prompted with a few-shot method to produce synthetic detoxification data. Then, the resulting data was employed in the prompt to GPT-3.5 to perform detoxification.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Team Yekaterina29 (mT5-XL)</head><p>The participant fine-tuned mT5-XL instances <ref type="bibr" target="#b50">[51]</ref> on the provided development multilingual parallel dataset. Almost all of the participant used the current state-of-the-art Large Language Models (LLMs), among which are GPT-3.5 <ref type="bibr" target="#b57">[58]</ref> and LLaMa-3 <ref type="bibr" target="#b70">[71]</ref> models. To enhance the model's performance on the task of detoxification, most participants used the few-shot prompting method. Among smaller models, mT5 <ref type="bibr" target="#b50">[51]</ref> and mT0 <ref type="bibr" target="#b66">[67]</ref> were utilized: usually, these models were fine-tuned using ad hoc filtering and data augmentation techniques, for instance, as RAG and backtranslation. Additionally, region-specific LLMs were also employed-Cotype <ref type="bibr" target="#b77">[78]</ref> and Kimi.AI <ref type="bibr" target="#b55">[56]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="8.">Results</head><p>Here, we provide the final results of the final test phase, of our tasks. The full detailed tables of results per each language and per each metric can be found in Appendix A.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="8.1.">Automatic Evaluation Leaderboard</head><p>We received 20 submissions for the development phase automatic leaderboard and 31 submissions for the test phase automatic leaderboard. Automatic evaluation leaderboards are publicly available online. <ref type="foot" target="#foot_19">20</ref>The final leaderboard from the test automatic phase evaluation is presented in Table <ref type="table">3</ref>.</p><p>The leading solutions were consistent across most languages, except for Spanish, Chinese, and Hindi. However, with the automatic evaluation leaderboard publicly available to all participants, some teams focused on optimizing their models specifically for the evaluation metrics, leading to potential overfitting.</p><p>Most solutions surpassed the baseline for at least one language, and in some cases, participant systems approached the performance of human references. However, except for Hindi, no participant solution outperformed human references in the automatic evaluation across any language. Although the automatic evaluation scores for human references across most languages hovered around a J score of 0.7, the results for Chinese were notably poor, with the highest participant score being 0.178 and the best human reference score at 0.201. The results leads to a further investigations of the robustness of the automatic evaluation metrics.</p><p>The top three teams across the majority of the languages generally employed a similar strategy, fine-tuning the mT0-XL text-to-text model. Team SmurfCat is holding the best automatic evaluation scores for all the languages, which was achieved by additionally fine-tuning mT0-XL with a recent ORPO alignment method. The majority of the submissions were multilingual, designed to cover all languages within a single model. These models demonstrated consistent score distributions across languages, with notable declines in performance for Chinese and Hindi. An exception was user ansafronov, who achieved the top score specifically for Chinese.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 3</head><p>Results of the automatic evaluation of the test phase. Scores are sorted by the average Joint score. The scores for each language are respective Joint scores. Baselines are highlighted with gray , Human References are highlighted with green . Three best scores for each language are highlighted with bold, the best score is underlined bold. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="8.2.">Human Evaluation Leaderboard</head><p>After participants confirmed their submissions via a form, we received 17 entries for the human evaluation phase. This evaluation was conducted on a subsample of 100 test set items through crowdsourcing.</p><p>The results of the human evaluation, organized by team and language, are publicly available. 21 The final leaderboard based on human evaluation is presented in Table <ref type="table">4</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 4</head><p>Results of the human final evaluation of the test phase. Scores are sorted by the average Joint score. The scores for each language are respective Joint scores. Baselines are highlighted with gray , Human References are highlighted with green . Three best scores for each language are highlighted with bold, the best score is underlined bold. The human evaluation leaderboard saw significant changes compared to the automatic evaluation phase. Human references received high scores from the annotators, with J scores around 0.8 or higher. However, not all teams surpassed the mT5 and Delete baselines. Interestingly, the Delete baseline outperformed the mT5 text-to-text generation baseline in languages such as Arabic, Hindi, Ukrainian, Russian, and Amharic. This indicates that not all multilingual models are equally proficient in understanding and handling toxicity across different languages.</p><p>In the human evaluation phase, participants' solutions closely matched the human references, even surpassing the provided references from parallel datasets in some languages. The top solution, after manual evaluation, was presented by user SomethingAwful and was based on the "uncensored" LLaMa3-70B language model. Interestingly, SomethingAwful's solution did not achieve the highest scores across all nine languages but excelled in Spanish, German, and Russian. The leader of the automatic evaluation leaderboard, Team SmurfCat, secured second place. Participants nikita.sushko and VitalyProtasov switched places in the manual leaderboard.</p><p>Similar to the automatic leaderboard, human assessments revealed that certain models excelled in specific languages. For instance, nikita.sushko and VitalyProtasov achieved top results in Arabic and Hindi. Despite Team mkrisnai ranking 7th overall, their solution performed exceptionally well in English, Spanish, Russian, and Ukrainian. Additionally, Team Team cake secured the highest scores specifically for English and Chinese.</p><p>From the detailed results in Appendix A, it is evident that the solutions surpassed human references in English, Spanish, and German, often achieving near-perfect fluency. However, this success does not extend to other languages. These results highlight the impressive human-like text generation capabilities of modern LLMs, though they still struggle with handling toxicity and maintaining consistent controllable generation across languages. Future work should focus on developing more challenging tasks, particularly in cross-lingual contexts.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="9.">Conclusion</head><p>In Multilingual Text Detoxification task at PAN 2024, participants were tasked with transforming text style from toxic to non-toxic across nine languages: English, Spanish, German, Chinese, Arabic, Hindi, Ukrainian, Russian, and Amharic. The task was divided into two phases: (i) development phase focused on cross-lingual transfer approaches; (ii) test phase utilized parallel training data for all languages and encouraged multilingual solutions. Participants' submissions in both phases underwent evaluation using a set of automatic metrics, followed by human evaluation of the test subset to determine the final leaderboard rankings.</p><p>Participants employed modern state-of-the-art Large Language Models either by prompting them in few-shot formats or fine-tuning medium-sized models. For certain languages with sufficient training data, these models approached or even exceeded human reference provided in the shared task. However, this was primarily observed for resource-rich European languages. Opportunities for enhancement remain significant for less resource-rich languages and those with limited data, highlighting the need for further exploration in cross-lingual text detoxification and knowledge transfer.        </p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head></head><label></label><figDesc>**k is this about? А н**рена ты здесь это писал? Та н**уй ти мені впав, скотина ти така))) Was für ein besch**senes Jahr Este país se va a la m**rda ‫ﺷرا**ط‬ ‫ﯾﺎ‬ ‫ﺑﺟﻧﺎزﺗﮫ‬ ‫وﺗﻣﺷوا‬ ‫اﻟﻘﺗﯾل‬ ‫ﺗﻘﺗﻠوا‬ አንተ ቆሻሻ በዚህ ወቅት አይንህን ማየት አልፈልግም 卧槽，抓到了！ ये माद**द डरे हु ए लग रहे है ? Detoxified Output What is this about? А зачем ты здесь это писал? Та навіщо ти мені потрібен Was für ein schlechtes Jahr. Cosas van muy mal en este país ‫ﺑﺟﻧﺎزﺗﮫ‬ ‫وﺗﻣﺷوا‬ ‫اﻟﻘﺗﯾل‬ ‫ﺗﻘﺗﻠوا‬ አንተ ጥሩ ሰው አይደለህም በዚህ ወቅት አንተን ማየት አልፈልግም 天啊，抓到了！ ये लोग डरे हु ए लग रहे है ?</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: In this work, we present a novel benchmark datasets for multilingual and cross-lingual text detoxification for 9 languages: English, Spanish, German, Chinese, Arabic, Hindi, Ukrainian, Russian, and Amharic.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>5/ 31 /</head><label>31</label><figDesc>24, 11:17 PM Toloka: Data solution to drive AI General tasks Which text is more offensive? -golden pairs Public description: Indicate which text is more offensive Private comment: toxic_pairwise Labeling completed! Start date: May 10, 2024 9:34 AM Accepted 100% (38 of 38) time per task suite -20 If number of fast task suite respo Skipped assignments If task suites skipped in a row ≥ Projects Users Skills Profile Messages Overview Efficiency Statistics Settings</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>1 2 3 I</head><label>13</label><figDesc>am Nancy and I like to ruin people's fun. 1) I am Nancy and I like to fuck up people's fun.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_5"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: The English interface example of crowdsourcing Task 1: pairwise toxicity comparison.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_6"><head></head><label></label><figDesc>For each Do these sentences mean the same? -golden pairs Public description: Check if the meanings of the two sentences are close or not. Private comment: content Labeling completed! Start date: May 20, 2024 3:07 PM Accepted 100% (39 of 39) time per task suite -15 R If number of fast task suite respons sentences mean the same thing? Dear Greenman, As usual you have proven to be an arse hole Dear Greenman, As usual, you have proven to be an unpleasant person Yes No still standing up for your lying crazy leader who loves himself and nobody else .</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_7"><head>Figure 3 :</head><label>3</label><figDesc>Figure 3: The English interface example of crowdsourcing Task 2: content similarity comparison.</figDesc><graphic coords="13,96.82,51.56,400.72,195.37" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_8"><head>Figure 4 :</head><label>4</label><figDesc>Figure 4: The English interface example of crowdsourcing Task 3: fluency evaluation.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Source # of Keywords</head><label></label><figDesc></figDesc><table><row><cell>English</cell><cell>[13, 47, 19]</cell><cell>3 390</cell></row><row><cell>Russian</cell><cell>[22, 19]</cell><cell>141 000</cell></row><row><cell>Ukrainian</cell><cell>[48, 19]</cell><cell>7 360</cell></row><row><cell>Spanish</cell><cell>[19]</cell><cell>1 200</cell></row><row><cell>German</cell><cell>[49, 19]</cell><cell>247</cell></row><row><cell>Hindi</cell><cell>[19]</cell><cell>133</cell></row><row><cell>Amharic</cell><cell>Ours+[19]</cell><cell>245</cell></row><row><cell>Arabic</cell><cell>Ours+[19]</cell><cell>430</cell></row><row><cell>Chinese</cell><cell>[50, 45, 19]</cell><cell>3 840</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_4"><head>Table 6</head><label>6</label><figDesc>Automatic and human evaluation results for Spanish.</figDesc><table><row><cell></cell><cell cols="3">Automatic Evaluation</cell><cell></cell><cell></cell><cell cols="2">Human Evaluation</cell></row><row><cell></cell><cell>STA</cell><cell>SIM</cell><cell>ChrF</cell><cell>J</cell><cell>STA</cell><cell>SIM</cell><cell>FL</cell><cell>J *</cell></row><row><cell>SomethingAwful</cell><cell cols="8">0.885 0.830 0.625 0.475 0.916 0.910 1.000 0.834</cell></row><row><cell>mkrisnai</cell><cell cols="8">0.867 0.806 0.584 0.421 0.886 0.940 1.000 0.833</cell></row><row><cell>VitalyProtasov</cell><cell cols="8">0.892 0.835 0.619 0.472 0.910 0.890 1.000 0.809</cell></row><row><cell>Human References</cell><cell cols="8">0.875 0.811 1.000 0.708 0.901 0.890 0.990 0.794</cell></row><row><cell>Team cake</cell><cell cols="8">0.928 0.765 0.488 0.360 0.891 0.870 0.990 0.767</cell></row><row><cell>Team NLPunks</cell><cell cols="8">0.861 0.848 0.615 0.458 0.906 0.860 0.980 0.764</cell></row><row><cell>Team SmurfCat</cell><cell cols="8">0.959 0.885 0.644 0.562 0.871 0.850 0.980 0.726</cell></row><row><cell>erehulka</cell><cell cols="8">0.884 0.865 0.634 0.496 0.930 0.770 0.990 0.708</cell></row><row><cell>Team SINAI</cell><cell cols="8">0.899 0.781 0.546 0.404 0.851 0.800 1.000 0.681</cell></row><row><cell>gleb.shnshn</cell><cell cols="8">0.900 0.799 0.584 0.436 0.890 0.760 1.000 0.676</cell></row><row><cell>Yekaterina29</cell><cell cols="8">0.745 0.888 0.646 0.439 0.835 0.760 1.000 0.634</cell></row><row><cell>nikita.sushko</cell><cell cols="8">0.788 0.896 0.657 0.480 0.866 0.720 0.990 0.617</cell></row><row><cell>Backtranslation</cell><cell cols="8">0.812 0.770 0.423 0.275 0.865 0.650 0.990 0.556</cell></row><row><cell>Delete</cell><cell cols="8">0.479 0.972 0.669 0.318 0.685 0.830 0.970 0.551</cell></row><row><cell cols="9">Team Iron Autobots 0.947 0.742 0.479 0.351 0.933 0.580 0.990 0.535</cell></row><row><cell>ZhongyuLuo</cell><cell cols="8">0.808 0.810 0.483 0.329 0.831 0.630 0.990 0.518</cell></row><row><cell>mT5</cell><cell cols="8">0.649 0.873 0.616 0.358 0.796 0.630 0.940 0.471</cell></row><row><cell>Team nlp_enjoyers</cell><cell cols="8">0.653 0.870 0.616 0.359 0.775 0.600 0.910 0.423</cell></row><row><cell>gangopsa</cell><cell cols="8">0.788 0.822 0.542 0.356 0.810 0.280 0.880 0.199</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_5"><head>Table 7</head><label>7</label><figDesc>Automatic and human evaluation results for German.</figDesc><table><row><cell></cell><cell cols="3">Automatic Evaluation</cell><cell></cell><cell></cell><cell cols="2">Human Evaluation</cell></row><row><cell></cell><cell>STA</cell><cell>SIM</cell><cell>ChrF</cell><cell>J</cell><cell>STA</cell><cell>SIM</cell><cell>FL</cell><cell>J *</cell></row><row><cell>SomethingAwful</cell><cell cols="8">0.799 0.904 0.759 0.550 0.898 0.990 1.000 0.889</cell></row><row><cell>erehulka</cell><cell cols="8">0.829 0.899 0.760 0.574 0.923 0.930 0.990 0.850</cell></row><row><cell>nikita.sushko</cell><cell cols="8">0.774 0.940 0.808 0.591 0.833 0.950 1.000 0.791</cell></row><row><cell>Team NLPunks</cell><cell cols="8">0.820 0.867 0.670 0.487 0.891 0.880 1.000 0.784</cell></row><row><cell>VitalyProtasov</cell><cell cols="8">0.646 0.951 0.813 0.502 0.798 0.980 0.990 0.774</cell></row><row><cell>Team cake</cell><cell cols="8">0.795 0.887 0.710 0.502 0.890 0.870 1.000 0.774</cell></row><row><cell>Yekaterina29</cell><cell cols="8">0.807 0.869 0.671 0.478 0.896 0.830 0.990 0.736</cell></row><row><cell>gangopsa</cell><cell cols="8">0.651 0.892 0.714 0.413 0.788 0.980 0.930 0.718</cell></row><row><cell>Human References</cell><cell cols="8">0.809 0.909 1.000 0.732 0.863 0.920 0.900 0.714</cell></row><row><cell>Team SmurfCat</cell><cell cols="8">0.921 0.923 0.781 0.677 0.856 0.830 0.980 0.696</cell></row><row><cell>mkrisnai</cell><cell cols="8">0.683 0.888 0.659 0.395 0.810 0.860 1.000 0.696</cell></row><row><cell cols="9">Team Iron Autobots 0.934 0.734 0.514 0.364 0.943 0.700 0.980 0.647</cell></row><row><cell>mT5</cell><cell cols="8">0.746 0.837 0.603 0.383 0.873 0.750 0.970 0.635</cell></row><row><cell>Delete</cell><cell cols="8">0.454 0.989 0.802 0.361 0.591 0.990 0.980 0.574</cell></row><row><cell>Team nlp_enjoyers</cell><cell cols="8">0.750 0.835 0.602 0.384 0.870 0.640 0.980 0.545</cell></row><row><cell>gleb.shnshn</cell><cell cols="8">0.910 0.803 0.617 0.464 0.940 0.580 1.000 0.545</cell></row><row><cell>Team SINAI</cell><cell cols="8">0.876 0.803 0.563 0.403 0.810 0.650 1.000 0.526</cell></row><row><cell>Backtranslation</cell><cell cols="8">0.796 0.747 0.372 0.232 0.858 0.400 1.000 0.343</cell></row><row><cell>ZhongyuLuo</cell><cell cols="8">0.815 0.222 0.130 0.024 0.876 0.010 0.990 0.008</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_6"><head>Table 8</head><label>8</label><figDesc>Automatic and human evaluation results for Chinese.</figDesc><table><row><cell></cell><cell cols="3">Automatic Evaluation</cell><cell></cell><cell></cell><cell cols="2">Human Evaluation</cell></row><row><cell></cell><cell>STA</cell><cell>SIM</cell><cell>ChrF</cell><cell>J</cell><cell>STA</cell><cell>SIM</cell><cell>FL</cell><cell>J *</cell></row><row><cell>Human References</cell><cell cols="8">0.266 0.789 1.000 0.201 0.963 0.990 0.970 0.925</cell></row><row><cell>Team cake</cell><cell cols="8">0.549 0.665 0.238 0.086 0.930 0.910 0.990 0.837</cell></row><row><cell>erehulka</cell><cell cols="8">0.389 0.789 0.551 0.160 0.950 0.870 0.820 0.677</cell></row><row><cell>Team NLPunks</cell><cell cols="8">0.462 0.815 0.395 0.150 0.648 0.980 0.950 0.603</cell></row><row><cell>Team SmurfCat</cell><cell cols="8">0.529 0.822 0.415 0.177 0.773 0.920 0.840 0.597</cell></row><row><cell>ZhongyuLuo</cell><cell cols="8">0.633 0.650 0.122 0.051 0.838 0.830 0.810 0.563</cell></row><row><cell>SomethingAwful</cell><cell cols="8">0.459 0.733 0.449 0.147 0.888 0.770 0.780 0.533</cell></row><row><cell cols="9">Team Iron Autobots 0.602 0.714 0.284 0.123 0.806 0.860 0.760 0.527</cell></row><row><cell>VitalyProtasov</cell><cell cols="8">0.411 0.868 0.504 0.175 0.891 0.970 0.570 0.493</cell></row><row><cell>nikita.sushko</cell><cell cols="8">0.415 0.869 0.504 0.176 0.920 0.990 0.520 0.473</cell></row><row><cell>mT5</cell><cell cols="8">0.289 0.809 0.411 0.095 0.726 0.920 0.650 0.434</cell></row><row><cell>Delete</cell><cell cols="8">0.384 0.887 0.524 0.174 0.693 0.990 0.620 0.425</cell></row><row><cell>gleb.shnshn</cell><cell cols="8">0.531 0.799 0.364 0.154 0.728 0.700 0.800 0.407</cell></row><row><cell>gangopsa</cell><cell cols="8">0.129 0.999 0.535 0.069 0.511 1.000 0.730 0.373</cell></row><row><cell>Backtranslation</cell><cell cols="8">0.661 0.591 0.070 0.026 0.831 0.600 0.690 0.344</cell></row><row><cell>mkrisnai</cell><cell cols="8">0.452 0.805 0.328 0.108 0.653 0.550 0.950 0.341</cell></row><row><cell>Team SINAI</cell><cell cols="8">0.608 0.741 0.286 0.126 0.558 0.720 0.830 0.333</cell></row><row><cell>Yekaterina29</cell><cell cols="8">0.344 0.778 0.472 0.130 0.840 0.830 0.430 0.299</cell></row><row><cell>Team nlp_enjoyers</cell><cell cols="8">0.375 0.770 0.403 0.104 0.778 0.430 0.690 0.230</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_7"><head>Table 9</head><label>9</label><figDesc>Automatic and human evaluation results for Arabic.</figDesc><table><row><cell></cell><cell cols="3">Automatic Evaluation</cell><cell></cell><cell></cell><cell cols="2">Human Evaluation</cell></row><row><cell></cell><cell>STA</cell><cell>SIM</cell><cell>ChrF</cell><cell>J</cell><cell>STA</cell><cell>SIM</cell><cell>FL</cell><cell>J *</cell></row><row><cell>nikita.sushko</cell><cell cols="8">0.780 0.930 0.783 0.575 0.921 0.990 0.970 0.885</cell></row><row><cell>Human References</cell><cell cols="8">0.795 0.875 1.000 0.694 0.941 0.920 0.950 0.823</cell></row><row><cell>Team SmurfCat</cell><cell cols="8">0.921 0.890 0.747 0.625 0.918 0.910 0.980 0.818</cell></row><row><cell>VitalyProtasov</cell><cell cols="8">0.730 0.921 0.775 0.522 0.891 0.930 0.950 0.787</cell></row><row><cell>erehulka</cell><cell cols="8">0.788 0.896 0.752 0.535 0.920 0.890 0.950 0.777</cell></row><row><cell>Team SINAI</cell><cell cols="8">0.883 0.699 0.425 0.282 0.921 0.830 1.000 0.764</cell></row><row><cell>SomethingAwful</cell><cell cols="8">0.825 0.860 0.719 0.513 0.931 0.820 0.970 0.741</cell></row><row><cell>Yekaterina29</cell><cell cols="8">0.695 0.904 0.710 0.452 0.828 0.850 1.000 0.704</cell></row><row><cell>Team NLPunks</cell><cell cols="8">0.728 0.857 0.652 0.414 0.866 0.840 0.950 0.691</cell></row><row><cell>Delete</cell><cell cols="8">0.597 0.974 0.777 0.455 0.750 0.920 0.940 0.648</cell></row><row><cell>mkrisnai</cell><cell cols="8">0.759 0.755 0.466 0.270 0.796 0.790 1.000 0.629</cell></row><row><cell>mT5</cell><cell cols="8">0.713 0.841 0.642 0.389 0.868 0.760 0.950 0.626</cell></row><row><cell cols="9">Team Iron Autobots 0.757 0.809 0.596 0.373 0.828 0.810 0.920 0.617</cell></row><row><cell>gangopsa</cell><cell cols="8">0.776 0.826 0.643 0.424 0.920 0.900 0.740 0.612</cell></row><row><cell>Team nlp_enjoyers</cell><cell cols="8">0.718 0.834 0.640 0.388 0.863 0.710 0.910 0.557</cell></row><row><cell>gleb.shnshn</cell><cell cols="8">0.794 0.825 0.616 0.415 0.920 0.650 0.910 0.544</cell></row><row><cell>ZhongyuLuo</cell><cell cols="8">0.771 0.719 0.366 0.225 0.832 0.590 0.990 0.486</cell></row><row><cell>Team cake</cell><cell cols="8">0.917 0.672 0.420 0.282 0.970 0.480 0.950 0.442</cell></row><row><cell>Backtranslation</cell><cell cols="8">0.836 0.682 0.319 0.205 0.915 0.460 0.990 0.416</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_8"><head>Table 10</head><label>10</label><figDesc>Automatic and human evaluation results for Hindi. .808 0.620 0.170 0.871 0.690 1.000 0.601 Team Iron Autobots 0.461 0.781 0.550 0.204 0.896 0.650 1.000 0.582TeamSINAI 0.586 0.750 0.490 0.224 0.960 0.570 0.990 0.541 erehulka 0.324 0.806 0.635 0.184 0.940 0.700 0.790 0.519 ZhongyuLuo 0.439 0.773 0.376 0.137 0.816 0.600 0.990 0.485 Team cake 0.771 0.583 0.310 0.157 0.953 0.360 0.990 0.339 Backtranslation 0.443 0.731 0.289 0.103 0.853 0.390 0.980 0.326</figDesc><table><row><cell></cell><cell cols="3">Automatic Evaluation</cell><cell></cell><cell></cell><cell cols="2">Human Evaluation</cell></row><row><cell></cell><cell>STA</cell><cell>SIM</cell><cell>ChrF</cell><cell>J</cell><cell>STA</cell><cell>SIM</cell><cell>FL</cell><cell>J *</cell></row><row><cell>Human References</cell><cell cols="8">0.367 0.814 1.000 0.297 0.975 0.990 1.000 0.965</cell></row><row><cell>VitalyProtasov</cell><cell cols="8">0.615 0.713 0.680 0.320 0.938 0.940 0.990 0.873</cell></row><row><cell>SomethingAwful</cell><cell cols="8">0.460 0.826 0.666 0.269 0.948 0.910 1.000 0.862</cell></row><row><cell>nikita.sushko</cell><cell cols="8">0.351 0.882 0.709 0.240 0.923 0.910 1.000 0.840</cell></row><row><cell>Team NLPunks</cell><cell cols="8">0.393 0.837 0.613 0.212 0.896 0.870 1.000 0.780</cell></row><row><cell>gangopsa</cell><cell cols="8">0.351 0.844 0.646 0.197 0.928 0.860 0.940 0.750</cell></row><row><cell>mkrisnai</cell><cell cols="8">0.476 0.786 0.509 0.193 0.871 0.840 1.000 0.732</cell></row><row><cell>Team SmurfCat</cell><cell cols="8">0.634 0.799 0.631 0.355 0.961 0.710 1.000 0.682</cell></row><row><cell>Team nlp_enjoyers</cell><cell cols="8">0.302 0.804 0.619 0.171 0.905 0.800 0.920 0.666</cell></row><row><cell>Yekaterina29</cell><cell cols="8">0.261 0.905 0.662 0.173 0.790 0.840 1.000 0.663</cell></row><row><cell>Delete</cell><cell cols="8">0.146 0.974 0.706 0.104 0.673 0.970 1.000 0.653</cell></row><row><cell>gleb.shnshn</cell><cell cols="8">0.497 0.790 0.595 0.244 0.975 0.670 0.990 0.646</cell></row><row><cell>mT5</cell><cell cols="2">0.295 0</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_9"><head>Table 11</head><label>11</label><figDesc>Automatic and human evaluation results for Ukrainian.</figDesc><table><row><cell></cell><cell cols="3">Automatic Evaluation</cell><cell></cell><cell></cell><cell cols="2">Human Evaluation</cell></row><row><cell></cell><cell>STA</cell><cell>SIM</cell><cell>ChrF</cell><cell>J</cell><cell>STA</cell><cell>SIM</cell><cell>FL</cell><cell>J *</cell></row><row><cell>Human References</cell><cell cols="8">0.877 0.899 1.000 0.790 0.990 0.980 0.930 0.902</cell></row><row><cell>Team SmurfCat</cell><cell cols="8">0.951 0.913 0.780 0.691 0.971 0.900 0.960 0.839</cell></row><row><cell>mkrisnai</cell><cell cols="8">0.895 0.842 0.592 0.460 0.963 0.770 0.990 0.734</cell></row><row><cell>SomethingAwful</cell><cell cols="8">0.875 0.887 0.733 0.584 0.966 0.710 1.000 0.686</cell></row><row><cell>nikita.sushko</cell><cell cols="8">0.886 0.919 0.804 0.668 0.965 0.720 0.970 0.673</cell></row><row><cell>VitalyProtasov</cell><cell cols="8">0.846 0.922 0.792 0.628 0.956 0.710 0.980 0.665</cell></row><row><cell>Team SINAI</cell><cell cols="8">0.944 0.797 0.551 0.436 0.983 0.690 0.970 0.658</cell></row><row><cell>Yekaterina29</cell><cell cols="8">0.804 0.891 0.742 0.553 0.940 0.710 0.980 0.654</cell></row><row><cell>Team NLPunks</cell><cell cols="8">0.771 0.869 0.665 0.466 0.936 0.710 0.950 0.631</cell></row><row><cell>erehulka</cell><cell cols="8">0.882 0.899 0.743 0.602 0.975 0.670 0.960 0.627</cell></row><row><cell>Delete</cell><cell cols="8">0.423 0.974 0.791 0.327 0.708 0.870 0.970 0.597</cell></row><row><cell>Team cake</cell><cell cols="8">0.804 0.863 0.658 0.470 0.966 0.580 0.890 0.498</cell></row><row><cell>gangopsa</cell><cell cols="8">0.816 0.884 0.721 0.527 0.943 0.540 0.950 0.483</cell></row><row><cell cols="9">Team Iron Autobots 0.861 0.807 0.561 0.403 0.930 0.530 0.970 0.478</cell></row><row><cell>gleb.shnshn</cell><cell cols="8">0.857 0.826 0.634 0.460 0.936 0.500 0.930 0.435</cell></row><row><cell>Team nlp_enjoyers</cell><cell cols="8">0.704 0.856 0.678 0.431 0.905 0.490 0.950 0.421</cell></row><row><cell>ZhongyuLuo</cell><cell cols="8">0.884 0.773 0.385 0.283 0.966 0.440 0.980 0.416</cell></row><row><cell>mT5</cell><cell cols="8">0.704 0.858 0.679 0.433 0.911 0.480 0.950 0.415</cell></row><row><cell>Backtranslation</cell><cell cols="8">0.914 0.704 0.293 0.201 0.981 0.230 1.000 0.225</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_10"><head>Table 12</head><label>12</label><figDesc>Automatic and human evaluation results for Russian. Delete 0.372 0.971 0.708 0.254 0.743 0.750 0.880 0.490 Team Iron Autobots 0.907 0.776 0.506 0.367 0.965 0.490 0.950 0.449 mT5 0.762 0.844 0.638 0.431 0.955 0.440 0.950 0.399 dkenco 0.595 0.817 0.554 0.264 0.825 0.480 0.990 0.392 Backtranslation 0.903 0.697 0.328 0.222 0.970 0.230 0.990 0.220 gangopsa 0.414 0.575 0.345 0.090 0.905 0.180 0.020 0.003</figDesc><table><row><cell></cell><cell cols="3">Automatic Evaluation</cell><cell></cell><cell></cell><cell cols="2">Human Evaluation</cell></row><row><cell></cell><cell>STA</cell><cell>SIM</cell><cell>ChrF</cell><cell>J</cell><cell>STA</cell><cell>SIM</cell><cell>FL</cell><cell>J *</cell></row><row><cell>SomethingAwful</cell><cell cols="8">0.819 0.873 0.695 0.515 0.986 0.850 1.000 0.838</cell></row><row><cell>Human References</cell><cell cols="8">0.887 0.824 1.000 0.732 0.990 0.830 0.970 0.797</cell></row><row><cell>mkrisnai</cell><cell cols="8">0.758 0.825 0.600 0.382 0.901 0.870 1.000 0.784</cell></row><row><cell>Team SmurfCat</cell><cell cols="8">0.957 0.885 0.736 0.634 0.953 0.830 0.960 0.759</cell></row><row><cell>nikita.sushko</cell><cell cols="8">0.843 0.901 0.728 0.570 0.948 0.800 0.980 0.743</cell></row><row><cell>VitalyProtasov</cell><cell cols="8">0.807 0.893 0.731 0.542 0.933 0.810 0.970 0.733</cell></row><row><cell>Team cake</cell><cell cols="8">0.881 0.791 0.540 0.394 0.958 0.740 1.000 0.709</cell></row><row><cell>Team MarSanAI</cell><cell cols="8">0.779 0.878 0.723 0.507 0.916 0.800 0.960 0.704</cell></row><row><cell>Yekaterina29</cell><cell cols="8">0.811 0.875 0.689 0.507 0.953 0.760 0.970 0.702</cell></row><row><cell>ZhongyuLuo</cell><cell cols="8">0.812 0.863 0.705 0.507 0.958 0.770 0.920 0.678</cell></row><row><cell>SINAI</cell><cell cols="8">0.890 0.792 0.533 0.396 0.935 0.740 0.980 0.678</cell></row><row><cell>erehulka</cell><cell cols="8">0.858 0.868 0.686 0.528 0.975 0.690 0.960 0.645</cell></row><row><cell>gleb.shnshn</cell><cell cols="8">0.857 0.817 0.627 0.445 0.955 0.670 0.960 0.614</cell></row><row><cell>Team NLPunks</cell><cell cols="8">0.709 0.858 0.630 0.402 0.938 0.570 0.950 0.508</cell></row><row><cell>Team nlp_enjoyers</cell><cell cols="8">0.762 0.842 0.638 0.431 0.920 0.580 0.940 0.501</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_11"><head>Table 13</head><label>13</label><figDesc>Automatic and human evaluation results for Amharic. .588 0.234 0.096 0.778 0.520 0.360 0.145 Team Iron Autobots 0.672 0.355 0.118 0.057 0.845 0.250 0.310 0.065</figDesc><table><row><cell></cell><cell cols="3">Automatic Evaluation</cell><cell></cell><cell></cell><cell cols="2">Human Evaluation</cell></row><row><cell></cell><cell>STA</cell><cell>SIM</cell><cell>ChrF</cell><cell>J</cell><cell>STA</cell><cell>SIM</cell><cell>FL</cell><cell>J *</cell></row><row><cell>Human References</cell><cell cols="8">0.893 0.683 1.000 0.601 0.935 0.990 0.920 0.851</cell></row><row><cell>ZhongyuLuo</cell><cell cols="8">0.819 0.665 0.165 0.095 0.875 0.890 0.930 0.724</cell></row><row><cell>SomethingAwful</cell><cell cols="8">0.776 0.855 0.438 0.299 0.801 0.980 0.910 0.714</cell></row><row><cell>Team SmurfCat</cell><cell cols="8">0.900 0.888 0.456 0.378 0.768 1.000 0.930 0.714</cell></row><row><cell>Team nlp_enjoyers</cell><cell cols="8">0.837 0.640 0.269 0.157 0.863 0.940 0.860 0.697</cell></row><row><cell>erehulka</cell><cell cols="8">0.586 0.971 0.482 0.286 0.700 1.000 0.980 0.686</cell></row><row><cell>nikita.sushko</cell><cell cols="8">0.742 0.908 0.478 0.328 0.755 0.990 0.910 0.680</cell></row><row><cell>VitalyProtasov</cell><cell cols="8">0.754 0.872 0.458 0.310 0.786 0.950 0.910 0.680</cell></row><row><cell>Delete</cell><cell cols="8">0.539 0.979 0.486 0.269 0.661 1.000 0.950 0.628</cell></row><row><cell>gangopsa</cell><cell cols="8">0.584 0.956 0.478 0.280 0.690 0.990 0.900 0.614</cell></row><row><cell>Team cake</cell><cell cols="8">0.559 0.836 0.360 0.178 0.691 0.960 0.920 0.610</cell></row><row><cell>mT5</cell><cell cols="8">0.836 0.641 0.270 0.157 0.893 0.840 0.810 0.607</cell></row><row><cell>Yekaterina29</cell><cell cols="8">0.794 0.589 0.204 0.102 0.891 0.980 0.690 0.602</cell></row><row><cell>Team NLPunks</cell><cell cols="8">0.555 0.865 0.372 0.194 0.743 0.880 0.860 0.562</cell></row><row><cell>Backtranslation</cell><cell cols="8">0.819 0.618 0.135 0.075 0.856 0.690 0.920 0.543</cell></row><row><cell>mkrisnai</cell><cell cols="8">0.467 0.946 0.453 0.205 0.515 0.990 0.960 0.489</cell></row><row><cell>gleb.shnshn</cell><cell cols="8">0.649 0.725 0.298 0.146 0.805 0.960 0.610 0.471</cell></row><row><cell>Team SINAI</cell><cell cols="2">0.623 0</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">https://pan.webis.de/clef24/pan24-web/text-detoxification.html</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_1">https://codalab.lisn.upsaclay.fr/competitions/18243</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_2">https://github.com/textdetox/textdetox_clef_2024/tree/main/instructions/paradetox_collection</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_3">https://huggingface.co/textdetox</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_4">https://toloka.ai</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="6" xml:id="foot_5">https://github.com/davidjurgens/potato</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="7" xml:id="foot_6">https://huggingface.co/datasets/textdetox/multilingual_paradetox</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="8" xml:id="foot_7">https://huggingface.co/datasets/textdetox/multilingual_paradetox_test</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="9" xml:id="foot_8">https://github.com/pan-webis-de/pan-code/tree/master/clef24/text-detoxification/baselines</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="10" xml:id="foot_9">huggingface.co/datasets/textdetox/multilingual_toxic_lexicon</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="11" xml:id="foot_10">https://huggingface.co/facebook/nllb-200-distilled-600M</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="12" xml:id="foot_11">https://huggingface.co/s-nlp/bart-base-detox</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="13" xml:id="foot_12">https://huggingface.co/google/mt5-xl</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="14" xml:id="foot_13">https://github.com/pan-webis-de/pan-code/blob/master/clef24/text-detoxification/evaluate.py</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="15" xml:id="foot_14">https://huggingface.co/datasets/textdetox/multilingual_toxicity_dataset</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="16" xml:id="foot_15">https://huggingface.co/textdetox/xlmr-large-toxicity-classifier</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="17" xml:id="foot_16">huggingface.co/sentence-transformers/LaBSE</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="18" xml:id="foot_17">https://github.com/textdetox/textdetox_clef_2024/tree/main/instructions/human_evaluation</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="19" xml:id="foot_18">https://toloka.ai</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="20" xml:id="foot_19">https://codalab.lisn.upsaclay.fr/competitions/18243#results</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgment</head><p>We express our deepest gratitude to Toloka platform to support our shared task. Crowdsourced data collection and human evaluation were made possible through the provided research grant.</p></div>
			</div>


			<div type="availability">
<div xmlns="http://www.tei-c.org/ns/1.0"><p>A. Panchenko) https://dardem.github.io (D. Dementieva); https://www.researchgate.net/profile/Daniil-Moskovskiy (D. Moskovskiy); https://github.com/bbkjunior/bbkjunior (N. Babakov); https://scholar.google.com/citations?user=g2m1wH4AAAAJ&amp;hl=en (A. A. Ayele); https://www.linkedin.com/in/naquee-rizwan-a97abb159 (N. Rizwan); https://www.linkedin.com/in/flo-schneider-hh (F. Schneider); https://ethanscuter.github.io (X. Wang); https://seyyaw.github.io (S. M. Yimam); https://linkedin.com/in/ustalov (D. Ustalov); https://github.com/eistakovskii (E. Stakovskii); https://www.sharjah.ac.ae/en/academics/Colleges/CI/dept/cs/Pages/ppl_detail.aspx?mcid=4 (A. Elnagar); https://cse.iitkgp.ac.in/~animeshm (A. Mukherjee); https://alexanderpanchenko.github.io (A</p></div>
			</div>

			<div type="annex">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>A. Automatic and Manual Evaluation Results per Language</head><p>Here, we provide the extended results-from both automatic and human evaluation setups-based on three evaluation parameters for all languages: English (Table <ref type="table">5</ref>), Spanish (Table <ref type="table">6</ref>), German (Table <ref type="table">7</ref>), Chinese (Table <ref type="table">8</ref>), Arabic (Table <ref type="table">9</ref>), Hindi (Table <ref type="table">10</ref>), Ukrainian (Table <ref type="table">11</ref>), Russian (Table <ref type="table">12</ref>), and Amharic (Table <ref type="table">13</ref>). In every table, the baselines are highlighted with gray ; Human References are highlighted with green ; the ordering is made by J score from Human Evaluation results. The automatic evaluation is based on the full test set of 600 samples per language; human evaluation was performed on 100 set of the test set per language. </p></div>			</div>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<author>
			<persName><forename type="first">Z</forename><forename type="middle">R</forename><surname>Shi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Fang</surname></persName>
		</author>
		<idno>CoRR abs/2001.01818</idno>
		<ptr target="http://arxiv.org/abs/2001.01818.arXiv:2001.01818" />
		<title level="m">Artificial intelligence for social good: A survey</title>
				<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">LLM) security and privacy: The good, the bad, and the ugly</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Yao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Duan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Cai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Sun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Zhang</surname></persName>
		</author>
		<idno type="DOI">10.48550/ARXIV.2312.02003</idno>
		<idno type="arXiv">arXiv:2312.02003</idno>
		<ptr target="https://doi.org/10.48550/arXiv.2312.02003.doi:10.48550/ARXIV.2312.02003" />
	</analytic>
	<monogr>
		<title level="m">A survey on large language model (</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Algorithmic censorship by social platforms: Power and resistance</title>
		<author>
			<persName><forename type="first">J</forename><surname>Cobbe</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Philosophy &amp; Technology</title>
		<imprint>
			<biblScope unit="volume">34</biblScope>
			<biblScope unit="page" from="739" to="766" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Hatexplain: A benchmark dataset for explainable hate speech detection</title>
		<author>
			<persName><forename type="first">B</forename><surname>Mathew</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Saha</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">M</forename><surname>Yimam</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Biemann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Goyal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Mukherjee</surname></persName>
		</author>
		<idno type="DOI">10.1609/AAAI.V35I17.17745</idno>
		<ptr target="https://doi.org/10.1609/aaai.v35i17.17745.doi:10.1609/AAAI.V35I17.17745" />
	</analytic>
	<monogr>
		<title level="m">Thirty-Fifth AAAI Conference on Artificial Intelligence, AAAI 2021, Thirty-Third Conference on Innovative Applications of Artificial Intelligence, IAAI 2021, The Eleventh Symposium on Educational Advances in Artificial Intelligence, EAAI 2021, Virtual Event</title>
				<imprint>
			<publisher>AAAI Press</publisher>
			<date type="published" when="2021">February 2-9, 2021. 2021</date>
			<biblScope unit="page" from="14867" to="14875" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Offensive language detection in spanish social media: Testing from bag-of-words to transformers models</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">M</forename><surname>Molero</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Pérez-Martín</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Á</forename><surname>Rodrigo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Peñas</surname></persName>
		</author>
		<idno type="DOI">10.1109/ACCESS.2023.3310244</idno>
		<ptr target="https://doi.org/10.1109/ACCESS.2023.3310244.doi:10.1109/ACCESS.2023.3310244" />
	</analytic>
	<monogr>
		<title level="j">IEEE Access</title>
		<imprint>
			<biblScope unit="volume">11</biblScope>
			<biblScope unit="page" from="95639" to="95652" />
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Exploring Amharic hate speech data collection and classification approaches</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">A</forename><surname>Ayele</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">M</forename><surname>Yimam</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">D</forename><surname>Belay</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Asfaw</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Biemann</surname></persName>
		</author>
		<ptr target="https://aclanthology.org/2023.ranlp-1.6" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 14th International Conference on Recent Advances in Natural Language Processing</title>
				<editor>
			<persName><forename type="first">R</forename><surname>Mitkov</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><surname>Angelova</surname></persName>
		</editor>
		<meeting>the 14th International Conference on Recent Advances in Natural Language Processing<address><addrLine>Shoumen, Bulgaria, Varna, Bulgaria</addrLine></address></meeting>
		<imprint>
			<publisher>INCOMA Ltd</publisher>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="49" to="59" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">A dataset of Hindi-English codemixed social media text for hate speech detection</title>
		<author>
			<persName><forename type="first">A</forename><surname>Bohra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Vijay</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Singh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">S</forename><surname>Akhtar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Shrivastava</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/W18-1105</idno>
		<ptr target="https://aclanthology.org/W18-1105.doi:10.18653/v1/W18-1105" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Second Workshop on Computational Modeling of People&apos;s Opinions, Personality, and Emotions in Social Media, Association for Computational Linguistics</title>
				<meeting>the Second Workshop on Computational Modeling of People&apos;s Opinions, Personality, and Emotions in Social Media, Association for Computational Linguistics<address><addrLine>New Orleans, Louisiana, USA</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page" from="36" to="41" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<monogr>
		<title level="m" type="main">Mutox: Universal multilingual audio-based toxicity dataset and zero-shot detector</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">R</forename><surname>Costa-Jussà</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">C</forename><surname>Meglioli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Andrews</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Dale</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Hansanti</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Kalbassi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Mourachko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Ropers</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Wood</surname></persName>
		</author>
		<idno type="DOI">10.48550/ARXIV.2401.05060</idno>
		<idno type="arXiv">arXiv:2401.05060</idno>
		<ptr target="https://doi.org/10.48550/arXiv.2401.05060.doi:10.48550/ARXIV.2401.05060" />
		<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Should democracies ban hate speech? hate speech laws and counterspeech</title>
		<author>
			<persName><forename type="first">E</forename><surname>Kulenović</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Ethical Theory and Moral Practice</title>
		<imprint>
			<biblScope unit="volume">26</biblScope>
			<biblScope unit="page" from="511" to="532" />
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Delete, retrieve, generate: a simple approach to sentiment and style transfer</title>
		<author>
			<persName><forename type="first">J</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Jia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>He</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Liang</surname></persName>
		</author>
		<idno type="DOI">10.18653/V1/N18-1169</idno>
		<ptr target="https://doi.org/10.18653/v1/n18-1169.doi:10.18653/V1/N18-1169" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2018</title>
				<editor>
			<persName><forename type="first">M</forename><forename type="middle">A</forename><surname>Walker</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">H</forename><surname>Ji</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Stent</surname></persName>
		</editor>
		<meeting>the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2018<address><addrLine>New Orleans, Louisiana, USA</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2018">June 1-6, 2018. 2018</date>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page" from="1865" to="1874" />
		</imprint>
	</monogr>
	<note>Association for Computational Linguistics</note>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Fighting offensive language on social media with unsupervised text style transfer</title>
		<author>
			<persName><forename type="first">C</forename><surname>Nogueira Dos Santos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Melnyk</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Padhi</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/P18-2031</idno>
		<ptr target="https://aclanthology.org/P18-2031.doi:10.18653/v1/P18-2031" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics</title>
				<meeting>the 56th Annual Meeting of the Association for Computational Linguistics<address><addrLine>Melbourne, Australia</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2018">2018</date>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="page" from="189" to="194" />
		</imprint>
	</monogr>
	<note>Short Papers), Association for Computational Linguistics</note>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Text detoxification using large pre-trained neural models</title>
		<author>
			<persName><forename type="first">D</forename><surname>Dale</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Voronov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Dementieva</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Logacheva</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Kozlova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Semenov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Panchenko</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/2021.emnlp-main.629</idno>
		<ptr target="https://aclanthology.org/2021.emnlp-main.629.doi:10.18653/v1/2021.emnlp-main.629" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, Association for Computational Linguistics, Online and</title>
				<meeting>the 2021 Conference on Empirical Methods in Natural Language Processing, Association for Computational Linguistics, Online and<address><addrLine>Punta Cana, Dominican Republic</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2021">2021</date>
			<biblScope unit="page" from="7979" to="7996" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">ParaDetox: Detoxification with parallel data</title>
		<author>
			<persName><forename type="first">V</forename><surname>Logacheva</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Dementieva</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Ustyantsev</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Moskovskiy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Dale</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Krotova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Semenov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Panchenko</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/2022.acl-long.469</idno>
		<ptr target="https://aclanthology.org/2022.acl-long.469.doi:10.18653/v1/2022.acl-long.469" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics</title>
		<title level="s">Long Papers</title>
		<meeting>the 60th Annual Meeting of the Association for Computational Linguistics<address><addrLine>Dublin, Ireland</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2022">2022</date>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page" from="6804" to="6818" />
		</imprint>
	</monogr>
	<note>Association for Computational Linguistics</note>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">APPDIA: A discourse-aware transformer-based style transfer model for offensive social media conversations</title>
		<author>
			<persName><forename type="first">K</forename><surname>Atwell</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Hassan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Alikhani</surname></persName>
		</author>
		<ptr target="https://aclanthology.org/2022.coling-1.530" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 29th International Conference on Computational Linguistics, COLING 2022</title>
				<editor>
			<persName><forename type="first">N</forename><surname>Calzolari</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">C</forename><surname>Huang</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">H</forename><surname>Kim</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Pustejovsky</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">L</forename><surname>Wanner</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">K</forename><surname>Choi</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Ryu</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">H</forename><surname>Chen</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">L</forename><surname>Donatelli</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">H</forename><surname>Ji</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S</forename><surname>Kurohashi</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Paggio</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Xue</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S</forename><surname>Kim</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Y</forename><surname>Hahm</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Z</forename><surname>He</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">T</forename><forename type="middle">K</forename><surname>Lee</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">E</forename><surname>Santus</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">F</forename><surname>Bond</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S</forename><surname>Na</surname></persName>
		</editor>
		<meeting>the 29th International Conference on Computational Linguistics, COLING 2022<address><addrLine>Gyeongju, Republic of Korea</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2022">October 12-17, 2022. 2022</date>
			<biblScope unit="page" from="6063" to="6074" />
		</imprint>
	</monogr>
	<note>International Committee on Computational Linguistics</note>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Exploring methods for cross-lingual text style transfer: The case of text detoxification</title>
		<author>
			<persName><forename type="first">D</forename><surname>Dementieva</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Moskovskiy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Dale</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Panchenko</surname></persName>
		</author>
		<idno type="DOI">10.18653/V1/2023.IJCNLP-MAIN.70</idno>
		<ptr target="https://doi.org/10.18653/v1/2023.ijcnlp-main.70.doi:10.18653/V1/2023.IJCNLP-MAIN.70" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 13th International Joint Conference on Natural Language Processing and the 3rd Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics, IJCNLP 2023 -Volume 1: Long Papers</title>
				<editor>
			<persName><forename type="first">J</forename><forename type="middle">C</forename><surname>Park</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Y</forename><surname>Arase</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">B</forename><surname>Hu</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">W</forename><surname>Lu</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">D</forename><surname>Wijaya</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Purwarianti</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><forename type="middle">A</forename><surname>Krisnadhi</surname></persName>
		</editor>
		<meeting>the 13th International Joint Conference on Natural Language Processing and the 3rd Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics, IJCNLP 2023 -Volume 1: Long Papers<address><addrLine>Nusa Dua, Bali</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2023">November 1 -4, 2023. 2023</date>
			<biblScope unit="page" from="1083" to="1101" />
		</imprint>
	</monogr>
	<note>Association for Computational Linguistics</note>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Overview of PAN 2024: Multi-Author Writing Style Analysis, Multilingual Text Detoxification, Oppositional Thinking Analysis, and Generative AI Authorship Verification</title>
		<author>
			<persName><forename type="first">J</forename><surname>Bevendorff</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><forename type="middle">B</forename><surname>Casals</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Chulvi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Dementieva</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Elnagar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Freitag</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Fröbe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Korenčić</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Mayerl</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Mukherjee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Panchenko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Potthast</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Rangel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Rosso</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Smirnova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Stamatatos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Stein</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Taulé</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Ustalov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Wiegmann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Zangerle</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Experimental IR Meets Multilinguality, Multimodality, and Interaction. Proceedings of the Fourteenth International Conference of the CLEF Association (CLEF 2024)</title>
		<title level="s">Lecture Notes in Computer Science</title>
		<meeting><address><addrLine>Berlin Heidelberg New York</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Six attributes of unhealthy conversations</title>
		<author>
			<persName><forename type="first">I</forename><surname>Price</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Gifford-Moore</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Flemming</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Musker</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Roichman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Sylvain</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Thain</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Dixon</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Sorensen</surname></persName>
		</author>
		<idno type="DOI">10.18653/V1/2020.ALW-1.15</idno>
		<ptr target="https://doi.org/10.18653/v1/2020.alw-1.15.doi:10.18653/V1/2020.ALW-1.15" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Fourth Workshop on Online Abuse and Harms, WOAH 2020</title>
				<editor>
			<persName><forename type="first">S</forename><surname>Akiwowo</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">B</forename><surname>Vidgen</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">V</forename><surname>Prabhakaran</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Z</forename><surname>Waseem</surname></persName>
		</editor>
		<meeting>the Fourth Workshop on Online Abuse and Harms, WOAH 2020</meeting>
		<imprint>
			<date type="published" when="2020-11-20">November 20, 2020. 2020</date>
			<biblScope unit="page" from="114" to="124" />
		</imprint>
	</monogr>
	<note>Association for Computational Linguistics</note>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Predicting different types of subtle toxicity in unhealthy online conversations</title>
		<author>
			<persName><forename type="first">S</forename><surname>Gilda</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Giovanini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Silva</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Oliveira</surname></persName>
		</author>
		<idno type="DOI">10.1016/J.PROCS.2021.12.254</idno>
		<ptr target="https://doi.org/10.1016/j.procs.2021.12.254.doi:10.1016/J.PROCS.2021.12.254" />
	</analytic>
	<monogr>
		<title level="m">The 12th International Conference on Emerging Ubiquitous Systems and Pervasive Networks (EUSPN 2021) / The 11th International Conference on Current and Future Trends of Information and Communication Technologies in Healthcare (ICTH-2021)</title>
		<title level="s">Procedia Computer Science</title>
		<editor>
			<persName><forename type="first">N</forename><surname>Varandas</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Yasar</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">H</forename><surname>Malik</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S</forename><surname>Galland</surname></persName>
		</editor>
		<meeting><address><addrLine>Leuven, Belgium</addrLine></address></meeting>
		<imprint>
			<publisher>Elsevier</publisher>
			<date type="published" when="2021">November 1-4, 2021. 2021</date>
			<biblScope unit="volume">198</biblScope>
			<biblScope unit="page" from="360" to="366" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<monogr>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">R</forename><surname>Costa-Jussà</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Cross</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Çelebi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Elbayad</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Heafield</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Heffernan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Kalbassi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Lam</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Licht</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Maillard</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">Y</forename><surname>Sun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Wenzek</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Youngblood</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Akula</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Barrault</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">M</forename><surname>Gonzalez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Hansanti</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Hoffman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Jarrett</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><forename type="middle">R</forename><surname>Sadagopan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Rowe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Spruit</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Tran</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Andrews</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><forename type="middle">F</forename><surname>Ayan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Bhosale</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Edunov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Fan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Gao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Goswami</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Guzmán</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Koehn</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Mourachko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Ropers</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Saleem</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Schwenk</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Wang</surname></persName>
		</author>
		<idno type="DOI">10.48550/ARXIV.2207.04672</idno>
		<idno type="arXiv">arXiv:2207.04672</idno>
		<ptr target="https://doi.org/10.48550/arXiv.2207.04672.doi:10.48550/ARXIV.2207.04672" />
		<title level="m">No language left behind: Scaling human-centered machine translation</title>
				<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">Methods for detoxification of texts for the russian language</title>
		<author>
			<persName><forename type="first">D</forename><surname>Dementieva</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Moskovskiy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Logacheva</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Dale</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Kozlova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Semenov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Panchenko</surname></persName>
		</author>
		<idno type="DOI">10.3390/MTI5090054</idno>
		<ptr target="https://doi.org/10.3390/mti5090054.doi:10.3390/MTI5090054" />
	</analytic>
	<monogr>
		<title level="j">Multimodal Technol. Interact</title>
		<imprint>
			<biblScope unit="volume">5</biblScope>
			<biblScope unit="page">54</biblScope>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">Exploring cross-lingual text detoxification with large multilingual language models</title>
		<author>
			<persName><forename type="first">D</forename><surname>Moskovskiy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Dementieva</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Panchenko</surname></persName>
		</author>
		<idno type="DOI">10.18653/V1/2022.ACL-SRW.26</idno>
		<ptr target="https://doi.org/10.18653/v1/2022.acl-srw.26.doi:10.18653/V1/2022.ACL-SRW.26" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics: Student Research Workshop, ACL 2022</title>
				<editor>
			<persName><forename type="first">S</forename><surname>Louvan</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Madotto</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">B</forename><surname>Madureira</surname></persName>
		</editor>
		<meeting>the 60th Annual Meeting of the Association for Computational Linguistics: Student Research Workshop, ACL 2022<address><addrLine>Dublin, Ireland</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2022">May 22-27, 2022. 2022</date>
			<biblScope unit="page" from="346" to="354" />
		</imprint>
	</monogr>
	<note>Association for Computational Linguistics</note>
</biblStruct>

<biblStruct xml:id="b21">
	<monogr>
		<author>
			<persName><forename type="first">D</forename><surname>Dementieva</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Logacheva</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Nikishina</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Fenogenova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Dale</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Krotova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Semenov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Shavrina</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Panchenko</surname></persName>
		</author>
		<author>
			<persName><surname>Russe-</surname></persName>
		</author>
		<ptr target="https://api.semanticscholar.org/CorpusID:253169495" />
		<title level="m">Findings of the First Russian Detoxification Shared Task Based on Parallel Corpora</title>
				<imprint>
			<date type="published" when="2022">2022. 2022</date>
		</imprint>
	</monogr>
	<note>COMPUTATIONAL LINGUISTICS AND INTELLECTUAL TECHNOLOGIES</note>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">Codalab competitions: An open source platform to organize scientific challenges</title>
		<author>
			<persName><forename type="first">A</forename><surname>Pavao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Guyon</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A.-C</forename><surname>Letournel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D.-T</forename><surname>Tran</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Baro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><forename type="middle">J</forename><surname>Escalante</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Escalera</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Thomas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Xu</surname></persName>
		</author>
		<ptr target="http://jmlr.org/papers/v24/21-1436.html" />
	</analytic>
	<monogr>
		<title level="j">Journal of Machine Learning Research</title>
		<imprint>
			<biblScope unit="volume">24</biblScope>
			<biblScope unit="page" from="1" to="6" />
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">Continuous Integration for Reproducible Shared Tasks with TIRA</title>
		<author>
			<persName><forename type="first">M</forename><surname>Fröbe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Wiegmann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Kolyada</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Grahm</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Elstner</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Loebe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Hagen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Stein</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Potthast</surname></persName>
		</author>
		<idno type="DOI">10.1007/978-3-031-28241-6_20</idno>
		<idno>doi:</idno>
		<ptr target="10.1007/978-3-031-28241-6_20" />
	</analytic>
	<monogr>
		<title level="m">Advances in Information Retrieval. 45th European Conference on IR Research (ECIR 2023)</title>
		<title level="s">Lecture Notes in Computer Science</title>
		<editor>
			<persName><forename type="first">J</forename><surname>Kamps</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">L</forename><surname>Goeuriot</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">F</forename><surname>Crestani</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><surname>Maistro</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">H</forename><surname>Joho</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">B</forename><surname>Davis</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">C</forename><surname>Gurrin</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">U</forename><surname>Kruschwitz</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Caputo</surname></persName>
		</editor>
		<meeting><address><addrLine>Berlin Heidelberg New York</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="236" to="241" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<analytic>
		<title level="a" type="main">A study on manual and automatic evaluation for text style transfer: The case of detoxification</title>
		<author>
			<persName><forename type="first">V</forename><surname>Logacheva</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Dementieva</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Krotova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Fenogenova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Nikishina</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Shavrina</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Panchenko</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/2022.humeval-1.8</idno>
		<ptr target="https://aclanthology.org/2022.humeval-1.8.doi:10.18653/v1/2022.humeval-1.8" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2nd Workshop on Human Evaluation of NLP Systems (HumEval), Association for Computational Linguistics</title>
				<meeting>the 2nd Workshop on Human Evaluation of NLP Systems (HumEval), Association for Computational Linguistics<address><addrLine>Dublin, Ireland</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2022">2022</date>
			<biblScope unit="page" from="90" to="101" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b25">
	<analytic>
		<title level="a" type="main">Paradetox: Detoxification with parallel data</title>
		<author>
			<persName><forename type="first">V</forename><surname>Logacheva</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Dementieva</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Ustyantsev</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Moskovskiy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Dale</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Krotova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Semenov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Panchenko</surname></persName>
		</author>
		<idno type="DOI">10.18653/V1/2022.ACL-LONG.469</idno>
		<ptr target="https://doi.org/10.18653/v1/2022.acl-long.469.doi:10.18653/V1/2022.ACL-LONG.469" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics</title>
				<editor>
			<persName><forename type="first">S</forename><surname>Muresan</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Nakov</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Villavicencio</surname></persName>
		</editor>
		<meeting>the 60th Annual Meeting of the Association for Computational Linguistics<address><addrLine>Dublin, Ireland</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2022">May 22-27, 2022. 2022</date>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page" from="6804" to="6818" />
		</imprint>
	</monogr>
	<note>Association for Computational Linguistics</note>
</biblStruct>

<biblStruct xml:id="b26">
	<monogr>
		<author>
			<persName><surname>Jigsaw</surname></persName>
		</author>
		<ptr target="https://www.kaggle.com/c/jigsaw-toxic-comment-classification-challenge" />
		<title level="m">Toxic comment classification challenge</title>
				<imprint>
			<date type="published" when="2017">2017. 2024-03-18</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b27">
	<monogr>
		<author>
			<persName><forename type="first">D</forename><surname>Dementieva</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Babakov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Panchenko</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2404.02037</idno>
		<title level="m">Multiparadetox: Extending text detoxification with parallel data to new languages</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b28">
	<monogr>
		<author>
			<persName><forename type="first">A</forename><surname>Belchikov</surname></persName>
		</author>
		<ptr target="https://www.kaggle.com/blackmoon/russian-language-toxic-comments" />
		<title level="m">Russian language toxic comments</title>
				<imprint>
			<date type="published" when="2019">2019. 2023-12-14</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b29">
	<monogr>
		<author>
			<persName><forename type="first">A</forename><surname>Semiletov</surname></persName>
		</author>
		<ptr target="https://www.kaggle.com/alexandersemiletov/toxic-russian-comments" />
		<title level="m">Toxic Russian Comments: Labelled comments from the popular Russian social network</title>
				<imprint>
			<date type="published" when="2020">2020. 2023-12-14</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b30">
	<monogr>
		<author>
			<persName><forename type="first">K</forename><surname>Bobrovnyk</surname></persName>
		</author>
		<ptr target="Ac-cessed:2023-12-14" />
		<title level="m">Ukrainian obscene lexicon</title>
				<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b31">
	<analytic>
		<title level="a" type="main">Automated building and analysis of ukrainian twitter corpus for toxic text detection</title>
		<author>
			<persName><forename type="first">K</forename><surname>Bobrovnyk</surname></persName>
		</author>
		<ptr target="https://ena.lpnu.ua:8443/server/api/core/bitstreams/c4c645c1-f465-4895-98dd-765f862cf186/content" />
	</analytic>
	<monogr>
		<title level="m">COLINS 2019</title>
				<imprint>
			<date type="published" when="2019">2019</date>
			<biblScope unit="volume">II</biblScope>
		</imprint>
	</monogr>
	<note>: Workshop</note>
</biblStruct>

<biblStruct xml:id="b32">
	<analytic>
		<title level="a" type="main">Detecting and monitoring hate speech in twitter</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">C</forename><surname>Pereira-Kohatsu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">Q</forename><surname>Sánchez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Liberatore</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Camacho-Collados</surname></persName>
		</author>
		<idno type="DOI">10.3390/S19214654</idno>
		<ptr target="https://doi.org/10.3390/s19214654.doi:10.3390/S19214654" />
	</analytic>
	<monogr>
		<title level="j">Sensors</title>
		<imprint>
			<biblScope unit="volume">19</biblScope>
			<biblScope unit="page">4654</biblScope>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b33">
	<monogr>
		<title level="m" type="main">Newscom-tox: a corpus of comments on news articles annotated for toxicity in spanish, Language Resources and Evaluation</title>
		<author>
			<persName><forename type="first">M</forename><surname>Taulé</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Nofre</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Bargiela</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Bonet</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2024">2024</date>
			<biblScope unit="page" from="1" to="41" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b34">
	<analytic>
		<title level="a" type="main">RoBERTuito: a pre-trained language model for social media text in Spanish</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">M</forename><surname>Pérez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">A</forename><surname>Furman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Alonso Alemany</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">M</forename><surname>Luque</surname></persName>
		</author>
		<ptr target="https://aclanthology.org/2022.lrec-1.785" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Thirteenth Language Resources and Evaluation Conference, European Language Resources Association</title>
				<meeting>the Thirteenth Language Resources and Evaluation Conference, European Language Resources Association<address><addrLine>Marseille, France</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2022">2022</date>
			<biblScope unit="page" from="7235" to="7243" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b35">
	<monogr>
		<author>
			<persName><forename type="first">M</forename><surname>Wiegand</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Siegel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Ruppenhofer</surname></persName>
		</author>
		<title level="m">Overview of the GermEval 2018 Shared Task on the Identification of Offensive Language</title>
				<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b36">
	<analytic>
		<title level="a" type="main">Overview of the GermEval 2021 Shared Task on the Identification of Toxic, Engaging, and Fact-Claiming Comments</title>
		<author>
			<persName><forename type="first">J</forename><surname>Risch</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Stoll</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Wilms</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Wiegand</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the GermEval 2021 Shared Task on the Identification of Toxic, Engaging, and Fact-Claiming Comments</title>
				<meeting>the GermEval 2021 Shared Task on the Identification of Toxic, Engaging, and Fact-Claiming Comments<address><addrLine>Duesseldorf, Germany</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2021">2021</date>
			<biblScope unit="page" from="1" to="12" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b37">
	<analytic>
		<title level="a" type="main">Measuring the Reliability of Hate Speech Annotations: The Case of the European Refugee Crisis</title>
		<author>
			<persName><forename type="first">B</forename><surname>Ross</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Rist</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Carbonell</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Cabrera</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Kurowsky</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Wojatzki</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of NLP4CMC III: 3rd Workshop on Natural Language Processing for Computer-Mediated Communication</title>
				<editor>
			<persName><forename type="first">M</forename><surname>Beißwenger</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><surname>Wojatzki</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">T</forename><surname>Zesch</surname></persName>
		</editor>
		<meeting>NLP4CMC III: 3rd Workshop on Natural Language Processing for Computer-Mediated Communication<address><addrLine>Bochum, Germany</addrLine></address></meeting>
		<imprint>
			<publisher>Bochumer Linguistische Arbeitsberichte</publisher>
			<date type="published" when="2016">2016</date>
			<biblScope unit="volume">17</biblScope>
			<biblScope unit="page">9</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b38">
	<analytic>
		<title level="a" type="main">Overview of the hasoc track at fire 2019: Hate speech and offensive content identification in indo-european languages</title>
		<author>
			<persName><forename type="first">T</forename><surname>Mandl</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Modha</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Majumder</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Patel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Dave</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Mandlia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Patel</surname></persName>
		</author>
		<idno type="DOI">10.1145/3368567.3368584</idno>
		<idno>doi:10.1145/3368567.3368584</idno>
		<ptr target="https://doi.org/10.1145/3368567.3368584" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 11th Annual Meeting of the Forum for Information Retrieval Evaluation, FIRE &apos;19</title>
				<meeting>the 11th Annual Meeting of the Forum for Information Retrieval Evaluation, FIRE &apos;19<address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>Association for Computing Machinery</publisher>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="14" to="17" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b39">
	<analytic>
		<title level="a" type="main">The 5Js in Ethiopia: Amharic hate speech data annotation using Toloka Crowdsourcing Platform</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">A</forename><surname>Ayele</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Dinter</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">D</forename><surname>Belay</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">T</forename><surname>Asfaw</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">M</forename><surname>Yimam</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Biemann</surname></persName>
		</author>
		<ptr target="https://ieeexplore.ieee.org/document/9971189" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 4th International Conference on Information and Communication Technology for Development for Africa (ICT4DA)</title>
				<meeting>the 4th International Conference on Information and Communication Technology for Development for Africa (ICT4DA)<address><addrLine>Bahir Dar, Ethiopia</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2022">2022</date>
			<biblScope unit="page" from="114" to="120" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b40">
	<analytic>
		<title level="a" type="main">L-hsab: A levantine twitter dataset for hate speech and abusive language</title>
		<author>
			<persName><forename type="first">H</forename><surname>Mulki</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Haddad</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">B</forename><surname>Ali</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Alshabani</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the third workshop on abusive language online</title>
				<meeting>the third workshop on abusive language online</meeting>
		<imprint>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="111" to="118" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b41">
	<analytic>
		<title level="a" type="main">T-hsab: A tunisian hate speech and abusive dataset</title>
		<author>
			<persName><forename type="first">H</forename><surname>Haddad</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Mulki</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Oueslati</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">International conference on Arabic language processing</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="251" to="263" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b42">
	<analytic>
		<title level="a" type="main">Overview of osact4 arabic offensive language detection shared task</title>
		<author>
			<persName><forename type="first">H</forename><surname>Mubarak</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Darwish</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Magdy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Elsayed</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Al-Khalifa</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 4th Workshop on open-source arabic corpora and processing tools, with a shared task on offensive language detection</title>
				<meeting>the 4th Workshop on open-source arabic corpora and processing tools, with a shared task on offensive language detection</meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="48" to="52" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b43">
	<analytic>
		<title level="a" type="main">Let-mi: An Arabic Levantine Twitter dataset for misogynistic language</title>
		<author>
			<persName><forename type="first">H</forename><surname>Mulki</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Ghanem</surname></persName>
		</author>
		<ptr target="https://aclanthology.org/2021.wanlp-1.16" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Sixth Arabic Natural Language Processing Workshop, Association for Computational Linguistics</title>
				<meeting>the Sixth Arabic Natural Language Processing Workshop, Association for Computational Linguistics<address><addrLine>Kyiv, Ukraine (Virtual</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2021">2021</date>
			<biblScope unit="page" from="154" to="163" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b44">
	<analytic>
		<title level="a" type="main">Facilitating fine-grained detection of Chinese toxic language: Hierarchical taxonomy, resources, and benchmarks</title>
		<author>
			<persName><forename type="first">J</forename><surname>Lu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Min</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Yang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Lin</surname></persName>
		</author>
		<ptr target="https://aclanthology.org/2023.acl-long.898" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics</title>
				<editor>
			<persName><forename type="first">A</forename><surname>Rogers</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Boyd-Graber</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Okazaki</surname></persName>
		</editor>
		<meeting>the 61st Annual Meeting of the Association for Computational Linguistics</meeting>
		<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="16235" to="16250" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b45">
	<analytic>
		<title level="a" type="main">Crowdsourcing of parallel corpora: the case of style transfer for detoxification</title>
		<author>
			<persName><forename type="first">D</forename><surname>Dementieva</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Ustyantsev</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Dale</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Kozlova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Semenov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Panchenko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Logacheva</surname></persName>
		</author>
		<ptr target="https://ceur-ws.org/Vol-2932/paper2.pdf" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2nd Crowd Science Workshop: Trust, Ethics, and Excellence in Crowdsourced Data Management at Scale colocated with 47th International Conference on Very Large Data Bases (VLDB 2021)</title>
		<title level="s">CEUR Workshop Proceedings</title>
		<editor>
			<persName><forename type="first">D</forename><surname>Ustalov</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">F</forename><surname>Casati</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Drutsa</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">I</forename><surname>Stelmakh</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Pavlichenko</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">D</forename><surname>Baidakova</surname></persName>
		</editor>
		<meeting>the 2nd Crowd Science Workshop: Trust, Ethics, and Excellence in Crowdsourced Data Management at Scale colocated with 47th International Conference on Very Large Data Bases (VLDB 2021)<address><addrLine>Copenhagen, Denmark</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2021-08-20">August 20, 2021. 2932. 2021</date>
			<biblScope unit="page" from="35" to="49" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b46">
	<monogr>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">J</forename><surname>Gabriel</surname></persName>
		</author>
		<ptr target="https://github.com/coffee-and-fun/google-profanity-words/blob/main/data/en.txt" />
		<title level="m">English full list of bad words and top swear words banned by google</title>
				<imprint>
			<date type="published" when="2023-12-12">2023. 2023-12-12</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b47">
	<monogr>
		<author>
			<persName><forename type="first">K</forename><surname>Bobrovnyk</surname></persName>
		</author>
		<ptr target="https://github.com/saganoren/obscene-ukr" />
		<title level="m">The dictionary of ukrainian obscene words</title>
				<imprint>
			<date type="published" when="2019">2019. 2023-12-12</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b48">
	<monogr>
		<author>
			<persName><forename type="first">I</forename><surname>Shutterstock</surname></persName>
		</author>
		<ptr target="https://github.com/LDNOOBW/List-of-Dirty-Naughty-Obscene-and-Otherwise-Bad-Words" />
		<title level="m">List of dirty, naughty, obscene, and otherwise bad words</title>
				<imprint>
			<date type="published" when="2020">2020. 2023-12-12</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b49">
	<analytic>
		<title level="a" type="main">SWSR: A chinese dataset and lexicon for online sexism detection</title>
		<author>
			<persName><forename type="first">A</forename><surname>Jiang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Yang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Zubiaga</surname></persName>
		</author>
		<idno type="DOI">10.1016/J.OSNEM.2021.100182</idno>
		<ptr target="https://doi.org/10.1016/j.osnem.2021.100182.doi:10.1016/J.OSNEM.2021.100182" />
	</analytic>
	<monogr>
		<title level="j">Online Soc. Networks Media</title>
		<imprint>
			<biblScope unit="volume">27</biblScope>
			<biblScope unit="page">100182</biblScope>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b50">
	<analytic>
		<title level="a" type="main">mt5: A massively multilingual pre-trained text-to-text transformer</title>
		<author>
			<persName><forename type="first">L</forename><surname>Xue</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Constant</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Roberts</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Kale</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Al-Rfou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Siddhant</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Barua</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Raffel</surname></persName>
		</author>
		<idno type="DOI">10.18653/V1/2021.NAACL-MAIN.41</idno>
		<ptr target="https://doi.org/10.18653/v1/2021.naacl-main.41.doi:10.18653/V1/2021.NAACL-MAIN.41" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2021, Online</title>
				<editor>
			<persName><forename type="first">K</forename><surname>Toutanova</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Rumshisky</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">L</forename><surname>Zettlemoyer</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">D</forename><surname>Hakkani-Tür</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">I</forename><surname>Beltagy</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S</forename><surname>Bethard</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">R</forename><surname>Cotterell</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">T</forename><surname>Chakraborty</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Y</forename><surname>Zhou</surname></persName>
		</editor>
		<meeting>the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2021, Online</meeting>
		<imprint>
			<date type="published" when="2021">June 6-11, 2021. 2021</date>
			<biblScope unit="page" from="483" to="498" />
		</imprint>
	</monogr>
	<note>Association for Computational Linguistics</note>
</biblStruct>

<biblStruct xml:id="b51">
	<analytic>
		<title level="a" type="main">Unsupervised cross-lingual representation learning at scale</title>
		<author>
			<persName><forename type="first">A</forename><surname>Conneau</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Khandelwal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Goyal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Chaudhary</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Wenzek</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Guzmán</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Grave</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Ott</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Zettlemoyer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Stoyanov</surname></persName>
		</author>
		<idno type="DOI">10.18653/V1/2020.ACL-MAIN.747</idno>
		<ptr target="https://doi.org/10.18653/v1/2020.acl-main.747.doi:10.18653/V1/2020.ACL-MAIN.747" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, ACL 2020</title>
				<editor>
			<persName><forename type="first">D</forename><surname>Jurafsky</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Chai</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Schluter</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><forename type="middle">R</forename><surname>Tetreault</surname></persName>
		</editor>
		<meeting>the 58th Annual Meeting of the Association for Computational Linguistics, ACL 2020</meeting>
		<imprint>
			<date type="published" when="2020">July 5-10, 2020. 2020</date>
			<biblScope unit="page" from="8440" to="8451" />
		</imprint>
	</monogr>
	<note>Association for Computational Linguistics</note>
</biblStruct>

<biblStruct xml:id="b52">
	<analytic>
		<title level="a" type="main">Language-agnostic BERT sentence embedding</title>
		<author>
			<persName><forename type="first">F</forename><surname>Feng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Yang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Cer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Arivazhagan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Wang</surname></persName>
		</author>
		<idno type="DOI">10.18653/V1/2022.ACL-LONG.62</idno>
		<ptr target="https://doi.org/10.18653/v1/2022.acl-long.62.doi:10.18653/V1/2022.ACL-LONG.62" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics</title>
				<editor>
			<persName><forename type="first">S</forename><surname>Muresan</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Nakov</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Villavicencio</surname></persName>
		</editor>
		<meeting>the 60th Annual Meeting of the Association for Computational Linguistics<address><addrLine>Dublin, Ireland</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2022">May 22-27, 2022. 2022</date>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page" from="878" to="891" />
		</imprint>
	</monogr>
	<note>Association for Computational Linguistics</note>
</biblStruct>

<biblStruct xml:id="b53">
	<analytic>
		<title level="a" type="main">A call for clarity in reporting BLEU scores</title>
		<author>
			<persName><forename type="first">M</forename><surname>Post</surname></persName>
		</author>
		<idno type="DOI">10.18653/V1/W18-6319</idno>
		<ptr target="https://doi.org/10.18653/v1/w18-6319.doi:10.18653/V1/W18-6319" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Third Conference on Machine Translation: Research Papers, WMT 2018</title>
				<editor>
			<persName><forename type="first">O</forename><surname>Bojar</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">R</forename><surname>Chatterjee</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">C</forename><surname>Federmann</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><surname>Fishel</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Y</forename><surname>Graham</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">B</forename><surname>Haddow</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><surname>Huck</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Jimeno-Yepes</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Koehn</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">C</forename><surname>Monz</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><surname>Negri</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Névéol</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><forename type="middle">L</forename><surname>Neves</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><surname>Post</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">L</forename><surname>Specia</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><surname>Turchi</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">K</forename><surname>Verspoor</surname></persName>
		</editor>
		<meeting>the Third Conference on Machine Translation: Research Papers, WMT 2018<address><addrLine>Belgium, Brussels</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2018-11-01">October 31 -November 1, 2018. 2018</date>
			<biblScope unit="page" from="186" to="191" />
		</imprint>
	</monogr>
	<note>Association for Computational Linguistics</note>
</biblStruct>

<biblStruct xml:id="b54">
	<analytic>
		<title level="a" type="main">A Multilingual Text Detoxification Method Based on Few-shot Learning and CO-STAR Framework</title>
		<author>
			<persName><forename type="first">J</forename><surname>Peng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Han</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Ye</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Guo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Lin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Tang</surname></persName>
		</author>
		<ptr target=".org" />
	</analytic>
	<monogr>
		<title level="m">Working Notes of CLEF 2024 -Conference and Labs of the Evaluation Forum</title>
				<editor>
			<persName><forename type="first">G</forename><surname>Faggioli</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Galuščáková</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><forename type="middle">G S</forename><surname>Herrera</surname></persName>
		</editor>
		<imprint>
			<publisher>CEUR-WS</publisher>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b55">
	<monogr>
		<author>
			<persName><forename type="first">M</forename><surname>Ai</surname></persName>
		</author>
		<ptr target="https://kimi.moonshot.cn" />
		<title level="m">Kimi chatbot</title>
				<imprint>
			<date type="published" when="2024-05-31">2024. 2024-05-31</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b56">
	<analytic>
		<title level="a" type="main">SINAI at PAN 2024 TextDetox: Application of Tree of Thought Strategy in Large Language Models for Multilingual Text Detoxification</title>
		<author>
			<persName><forename type="first">M</forename><surname>Vallecillo-Rodríguez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">M</forename><surname>Martín-Valdivia</surname></persName>
		</author>
		<ptr target=".org" />
	</analytic>
	<monogr>
		<title level="m">Working Notes of CLEF 2024 -Conference and Labs of the Evaluation Forum</title>
				<editor>
			<persName><forename type="first">G</forename><surname>Faggioli</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Galuščáková</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><forename type="middle">G S</forename><surname>Herrera</surname></persName>
		</editor>
		<imprint>
			<publisher>CEUR-WS</publisher>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b57">
	<monogr>
		<title level="m" type="main">Chatgpt: Optimizing language models for dialogue</title>
		<author>
			<persName><surname>Openai</surname></persName>
		</author>
		<ptr target="https://openai.com/blog/chatgpt" />
		<imprint>
			<date type="published" when="2022">2022. 2024-05-31</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b58">
	<analytic>
		<title level="a" type="main">TextDetox: ToxiCleanse RL and Paving the Way for Toxicity-Free Online Discourse</title>
		<author>
			<persName><forename type="first">M</forename><surname>Najafi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Tavan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Colreavy</surname></persName>
		</author>
		<author>
			<persName><surname>Marsan</surname></persName>
		</author>
		<author>
			<persName><surname>Pan</surname></persName>
		</author>
		<ptr target=".org" />
	</analytic>
	<monogr>
		<title level="m">Working Notes of CLEF 2024 -Conference and Labs of the Evaluation Forum</title>
				<editor>
			<persName><forename type="first">G</forename><surname>Faggioli</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Galuščáková</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><forename type="middle">G S</forename><surname>Herrera</surname></persName>
		</editor>
		<imprint>
			<publisher>CEUR-WS</publisher>
			<date type="published" when="2024">2024. 2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b59">
	<analytic>
		<title/>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">Q</forename><surname>Jiang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Sablayrolles</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Mensch</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Bamford</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">S</forename><surname>Chaplot</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>De Las Casas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Bressand</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Lengyel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Lample</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Saulnier</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">R</forename><surname>Lavaud</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Lachaux</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Stock</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">L</forename><surname>Scao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Lavril</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Lacroix</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><forename type="middle">E</forename><surname>Sayed</surname></persName>
		</author>
		<idno type="DOI">10.48550/ARXIV.2310.06825</idno>
		<idno type="arXiv">arXiv:2310.06825</idno>
		<ptr target="/ARXIV.2310.06825" />
	</analytic>
	<monogr>
		<title level="j">Mistral</title>
		<imprint>
			<biblScope unit="volume">7</biblScope>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b60">
	<monogr>
		<author>
			<persName><forename type="first">J</forename><surname>Schulman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Wolski</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Dhariwal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Radford</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Klimov</surname></persName>
		</author>
		<idno>CoRR abs/1707.06347</idno>
		<ptr target="http://arxiv.org/abs/1707.06347.arXiv:1707.06347" />
		<title level="m">Proximal policy optimization algorithms</title>
				<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b61">
	<monogr>
		<author>
			<persName><forename type="first">L</forename><surname>Werra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Belkada</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Tunstall</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Beeching</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Thrush</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Lambert</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Huang</surname></persName>
		</author>
		<ptr target="https://github.com/huggingface/trl" />
		<title level="m">Trl: Transformer reinforcement learning</title>
				<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b62">
	<analytic>
		<title level="a" type="main">HybridDetox: Combining Supervised and Unsupervised Methods for Effective Multilingual Text Detoxification</title>
		<author>
			<persName><forename type="first">S</forename><surname>Gangopadhyay</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Khan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Jabeen</surname></persName>
		</author>
		<ptr target=".org" />
	</analytic>
	<monogr>
		<title level="m">Working Notes of CLEF 2024 -Conference and Labs of the Evaluation Forum</title>
				<editor>
			<persName><forename type="first">G</forename><surname>Faggioli</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Galuščáková</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><forename type="middle">G S</forename><surname>Herrera</surname></persName>
		</editor>
		<imprint>
			<publisher>CEUR-WS</publisher>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b63">
	<analytic>
		<title level="a" type="main">Exploring the limits of transfer learning with a unified text-to-text transformer</title>
		<author>
			<persName><forename type="first">C</forename><surname>Raffel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Shazeer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Roberts</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Narang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Matena</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Zhou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">J</forename><surname>Liu</surname></persName>
		</author>
		<ptr target="http://jmlr.org/papers/v21/20-074.html" />
	</analytic>
	<monogr>
		<title level="j">J. Mach. Learn. Res</title>
		<imprint>
			<biblScope unit="volume">21</biblScope>
			<biblScope unit="page">67</biblScope>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b64">
	<analytic>
		<title level="a" type="main">BART: denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension</title>
		<author>
			<persName><forename type="first">M</forename><surname>Lewis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Goyal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Ghazvininejad</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Mohamed</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Levy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Stoyanov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Zettlemoyer</surname></persName>
		</author>
		<idno type="DOI">10.18653/V1/2020.ACL-MAIN.703</idno>
		<ptr target="https://doi.org/10.18653/v1/2020.acl-main.703.doi:10.18653/V1/2020.ACL-MAIN.703" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, ACL 2020</title>
				<editor>
			<persName><forename type="first">D</forename><surname>Jurafsky</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Chai</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Schluter</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><forename type="middle">R</forename><surname>Tetreault</surname></persName>
		</editor>
		<meeting>the 58th Annual Meeting of the Association for Computational Linguistics, ACL 2020</meeting>
		<imprint>
			<date type="published" when="2020">July 5-10, 2020. 2020</date>
			<biblScope unit="page" from="7871" to="7880" />
		</imprint>
	</monogr>
	<note>Association for Computational Linguistics</note>
</biblStruct>

<biblStruct xml:id="b65">
	<analytic>
		<title level="a" type="main">Multilingual TextDetox: Exploring Cross-lingual Transfer in Case of Large Language Models</title>
		<author>
			<persName><forename type="first">V</forename><surname>Protasov</surname></persName>
		</author>
		<author>
			<persName><surname>Pan</surname></persName>
		</author>
		<ptr target=".org" />
	</analytic>
	<monogr>
		<title level="m">Working Notes of CLEF 2024 -Conference and Labs of the Evaluation Forum</title>
				<editor>
			<persName><forename type="first">G</forename><surname>Faggioli</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Galuščáková</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><forename type="middle">G S</forename><surname>Herrera</surname></persName>
		</editor>
		<imprint>
			<publisher>CEUR-WS</publisher>
			<date type="published" when="2024">2024. 2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b66">
	<analytic>
		<title level="a" type="main">Crosslingual generalization through multitask finetuning</title>
		<author>
			<persName><forename type="first">N</forename><surname>Muennighoff</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Sutawika</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Roberts</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Biderman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">L</forename><surname>Scao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">S</forename><surname>Bari</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Shen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><forename type="middle">X</forename><surname>Yong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Schoelkopf</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Tang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Radev</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">F</forename><surname>Aji</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Almubarak</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Albanie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Alyafeai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Webson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Raff</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Raffel</surname></persName>
		</author>
		<idno type="DOI">10.18653/V1/2023.ACL-LONG.891</idno>
		<ptr target="https://doi.org/10.18653/v1/2023.acl-long.891.doi:10.18653/V1/2023.ACL-LONG.891" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics</title>
				<editor>
			<persName><forename type="first">A</forename><surname>Rogers</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><forename type="middle">L</forename><surname>Boyd-Graber</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Okazaki</surname></persName>
		</editor>
		<meeting>the 61st Annual Meeting of the Association for Computational Linguistics<address><addrLine>Toronto, Canada</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2023">July 9-14, 2023. 2023</date>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page" from="15991" to="16111" />
		</imprint>
	</monogr>
	<note>Association for Computational Linguistics</note>
</biblStruct>

<biblStruct xml:id="b67">
	<analytic>
		<title level="a" type="main">Multilingual TextDetox: Exploring Different Regimes For Synthetic Data Training For Multilingual Text Detoxification</title>
		<author>
			<persName><forename type="first">N</forename><surname>Sushko</surname></persName>
		</author>
		<author>
			<persName><surname>Pan</surname></persName>
		</author>
		<ptr target=".org" />
	</analytic>
	<monogr>
		<title level="m">Working Notes of CLEF 2024 -Conference and Labs of the Evaluation Forum</title>
				<editor>
			<persName><forename type="first">G</forename><surname>Faggioli</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Galuščáková</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><forename type="middle">G S</forename><surname>Herrera</surname></persName>
		</editor>
		<imprint>
			<publisher>CEUR-WS</publisher>
			<date type="published" when="2024">2024. 2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b68">
	<analytic>
		<title level="a" type="main">SmurfCat at PAN TexDetox 2024: Alignment of Multilingual Transformers for Text Detoxification</title>
		<author>
			<persName><forename type="first">E</forename><surname>Rykov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Zaytsev</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Anisimov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Voronin</surname></persName>
		</author>
		<ptr target=".org" />
	</analytic>
	<monogr>
		<title level="m">Working Notes of CLEF 2024 -Conference and Labs of the Evaluation Forum</title>
				<editor>
			<persName><forename type="first">G</forename><surname>Faggioli</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Galuščáková</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><forename type="middle">G S</forename><surname>Herrera</surname></persName>
		</editor>
		<imprint>
			<publisher>CEUR-WS</publisher>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b69">
	<monogr>
		<author>
			<persName><forename type="first">J</forename><surname>Hong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Thorne</surname></persName>
		</author>
		<idno type="DOI">10.48550/ARXIV.2403.07691</idno>
		<idno type="arXiv">arXiv:2403.07691</idno>
		<ptr target="/ARXIV.2403.07691" />
		<title level="m">ORPO: monolithic preference optimization without reference model</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b70">
	<monogr>
		<ptr target="https://github.com/meta-llama/llama3/blob/main/MODEL_CARD.md" />
		<title level="m">Llama 3 model card</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
	<note>AI@Meta</note>
</biblStruct>

<biblStruct xml:id="b71">
	<analytic>
		<title level="a" type="main">Memu_pro_kotow at PAN 2024 TextDetox: Uncensored Llama3 Helps to Censor Better</title>
		<author>
			<persName><forename type="first">S</forename><surname>Pletenev</surname></persName>
		</author>
		<ptr target=".org" />
	</analytic>
	<monogr>
		<title level="m">Working Notes of CLEF 2024 -Conference and Labs of the Evaluation Forum</title>
				<editor>
			<persName><forename type="first">G</forename><surname>Faggioli</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Galuščáková</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><forename type="middle">G S</forename><surname>Herrera</surname></persName>
		</editor>
		<imprint>
			<publisher>CEUR-WS</publisher>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b72">
	<monogr>
		<author>
			<persName><forename type="first">A</forename><surname>Zou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Phan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Campbell</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Guo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Ren</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Pan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Yin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Mazeika</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Dombrowski</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Goel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">J</forename><surname>Byun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Mallen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Basart</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Koyejo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Song</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Fredrikson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">Z</forename><surname>Kolter</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Hendrycks</surname></persName>
		</author>
		<idno type="DOI">10.48550/ARXIV.2310.01405</idno>
		<idno type="arXiv">arXiv:2310.01405</idno>
		<ptr target="https://doi.org/10.48550/arXiv.2310.01405.doi:10.48550/ARXIV.2310.01405" />
		<title level="m">Representation engineering: A top-down approach to AI transparency</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b73">
	<analytic>
		<title level="a" type="main">Multilingual Text Detoxification Using Google Cloud Translation and Post-Processing</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Luo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Luo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Wang</surname></persName>
		</author>
		<ptr target=".org" />
	</analytic>
	<monogr>
		<title level="m">Working Notes of CLEF 2024 -Conference and Labs of the Evaluation Forum</title>
				<editor>
			<persName><forename type="first">G</forename><surname>Faggioli</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Galuščáková</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><forename type="middle">G S</forename><surname>Herrera</surname></persName>
		</editor>
		<imprint>
			<publisher>CEUR-WS</publisher>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b74">
	<monogr>
		<title/>
		<author>
			<persName><forename type="first">A</forename><surname>Forever</surname></persName>
		</author>
		<ptr target="https://huggingface.co/ai-forever" />
		<imprint>
			<date type="published" when="2022">rugpt3. 2022. 2024-05-31</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b75">
	<analytic>
		<title level="a" type="main">nlp_enjoyers at Multilingual Textual Detoxification</title>
		<author>
			<persName><forename type="first">V</forename><surname>Zinkovich</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Karpukhin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Kurdiukov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Tikhomirov</surname></persName>
		</author>
		<ptr target=".org" />
	</analytic>
	<monogr>
		<title level="m">Working Notes of CLEF 2024 -Conference and Labs of the Evaluation Forum</title>
				<editor>
			<persName><forename type="first">G</forename><surname>Faggioli</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Galuščáková</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><forename type="middle">G S</forename><surname>Herrera</surname></persName>
		</editor>
		<imprint>
			<publisher>CEUR-WS</publisher>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
	<note>CLEF-2024</note>
</biblStruct>

<biblStruct xml:id="b76">
	<analytic>
		<title level="a" type="main">RAG Meets Detox: Enhancing Text Detoxification Using Open-Source Large Language Models with Retrieval Augmented Generation</title>
		<author>
			<persName><forename type="first">E</forename><surname>Řehulka</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Šuppa</surname></persName>
		</author>
		<ptr target=".org" />
	</analytic>
	<monogr>
		<title level="m">Working Notes of CLEF 2024 -Conference and Labs of the Evaluation Forum</title>
				<editor>
			<persName><forename type="first">G</forename><surname>Faggioli</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Galuščáková</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><forename type="middle">G S</forename><surname>Herrera</surname></persName>
		</editor>
		<imprint>
			<publisher>CEUR-WS</publisher>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b77">
	<monogr>
		<author>
			<persName><surname>Mts</surname></persName>
		</author>
		<author>
			<persName><surname>Ai</surname></persName>
		</author>
		<ptr target="https://mts.ai" />
		<title level="m">Cotype: Generative ai solutions</title>
				<imprint>
			<date type="published" when="2022">2022. 2024-05-31</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
