<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Toxicity Detection for Ukrainian-Language Texts in the TextAttributor System⋆</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Nataliia Darchuk</string-name>
          <email>n.darchuk@knu.ua</email>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Mykola Sazhok</string-name>
          <email>sazhok@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Institute for Information Technologies and Systems of the National Academy of Sciences of Ukraine</institution>
          ,
          <addr-line>40 prospekt Akademika Hlushkova, Kyiv, 03187</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Institute of the Ukrainian Language of the National Academy of Sciences of Ukraine</institution>
          ,
          <addr-line>4 Mykhailo Hrushevskyi Street, Kyiv, 01001</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Taras Shevchenko National University of Kyiv</institution>
          ,
          <addr-line>60 Volodymyrska Street, Kyiv, 01033</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
      </contrib-group>
      <fpage>68</fpage>
      <lpage>83</lpage>
      <abstract>
        <p>This paper presents the development and results of two distinct modules designed to detect toxic language in Ukrainian texts, primarily implemented within the "TextAttributor 1.0" expert system. The first module utilizes a rule-based approach, analyzing text through predefined linguistic rules and lexiconbased methods, while the second employs machine learning techniques, specifically leveraging the fastText and LLAMA-3 models, to automatically detect toxic content. The rule-based module outputs a detailed linguistic analysis, mapping toxic vocabulary using a precompiled lexicographic database, while the machine learning module calculates toxicity based on statistical models. The performance of both methods was evaluated by comparing their results on a corpus of Ukrainian texts, with the Pearson to effectively identify toxic content, contributing to ongoing efforts to mitigate the spread of harmful information. This paper contains rude texts that only serve as illustrative examples.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;toxic text detection</kwd>
        <kwd>sentiment analysis</kwd>
        <kwd>Ukrainian language</kwd>
        <kwd>lexicon-based method</kwd>
        <kwd>deep learning</kwd>
        <kwd>text classification 1</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        The study of text toxicity represents a relatively new field of inquiry within the broader domain of
sentiment analysis. The tonality of a text is a significant element that influences how it is perceived
and understood by the reader. It also enables the author to achieve their communicative objectives.
For this reason, the task of determining the tone of a text is of great interest today not only to
modern linguists and computer scientists but also to political scientists, managers, marketers,
advertisers, image makers, and other professionals working with a particular brand. This task has
become particularly crucial with the advent of the Internet, as it has provided a new platform for
the analysis of media texts that "transmit, store, and reproduce information that influences public
opinion" [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. The increased attention to the negative, toxic, and emotional component of textual
information occurred during the hybrid war period. The capacity of contemporary Internet media
and social networks to exert a deleterious influence on vast numbers of individuals through the
introduction of destructive and other harmful information into the mind or subconscious, which
      </p>
      <p>0000-0001-8932-9301 (N. Darchuk); 0000-0002-2644-3892 (O. Zuban); 0000-0003-2266-7650 (V. Robeiko);
0000-00029684-3840 (Yu. Tsyhvintseva); 0000-0003-1169-6851 (M. Sazhok)
leads to an inadequate perception of reality, underscores the acute urgency of the problem of
"ecology" and the protection of the Internet space.</p>
      <p>
        The issue of addressing the proliferation of destructive online content is a matter of global
concern and significance. Since 2019, June 18 has been designated by the United Nations as the
International Day against Hate Speech. This year, the Council of Europe held the Week against
Hate Speech on June 17-20 in Strasbourg. In 2022, the Committee of Ministers of the Council of
Europe developed Recommendations on combating hate speech [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. It is important to note that
there is no clear and universally accepted definition of toxic speech or hate speech. The definitions
of these concepts vary across international documents in the legislative, social, and linguistic fields.
However, they all have one thing in common: they describe texts that manifest aggression against
other people, nations, social groups, etc. and that violate human rights.
      </p>
      <p>The full-scale Russian-Ukrainian war has highlighted the urgency of this issue, as hate speech
and other forms of aggressive propaganda represent a crucial element of the Russian propaganda
apparatus in its efforts to erode the identity and integrity of the Ukrainian nation. To address this
issue, it is essential to develop tools for automated analysis and detection of Ukrainian-language
textual content that negatively impacts an individual's psychological state, public consciousness,
and infringes upon the rights and legitimate interests of users, society, and the state.</p>
      <p>
        In light of the pressing necessity for such tools, our team has developed IT solutions for the
automated identification of toxicity in Ukrainian-language text. These tools are integrated into
TextAttributor 1.0 [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], a linguistic parameterization expert system for Ukrainian-language media
texts. The research tasks were divided into two successive stages: (a) development of the module
for generating linguistic expertise of toxic text, detecting the toxicity index according to
dictionaries and rules and (b) forming a dataset of toxic texts using this module and human
expertise for machine learning. Additionally, a deep learning model was trained and evaluated by
the dataset, forming a machine learning module for toxicity detection. The concept of developing
an automated system for assessing the toxicity of text is rooted in the team's extensive research
experience in computational linguistics and a particular interest in sentiment analysis [
        <xref ref-type="bibr" rid="ref5 ref6 ref7 ref8 ref9">5, 6, 7, 8, 9</xref>
        ].
      </p>
      <p>The objective of the present article is twofold: firstly, to analyze the implementation of two
methods the lexicon and rules-based method, and the deep learning-based method in the
automatic analysis of the toxicity of Ukrainian-language texts; secondly, to investigate the
effectiveness of the TextAttributor 1.0 system using the two methods on the texts analyzed by the
system. The object of this study is Ukrainian-language texts. The subject of this study is the criteria
and methods used for the automatic identification of toxic Ukrainian-language texts. The study
employs a range of methods, including: a rules and lexicon-based sentiment analysis method, a
deep learning-based method, and a combination of linguistic methods, namely component analysis,
distributional analysis, and taxonomic analysis. Additionally, statistical methods, such as toxicity
indexing and the calculation of the Pearson correlation coefficient for sample data, are utilized.
Graphical modeling of statistical data is also employed as a method to visualize 520
Ukrainianlanguage texts that were submitted by users of the TextAttributor 1.0 web application.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Works</title>
      <p>
        The latest developments in the field indicate that the task of automatic toxicity detection is
currently solved mainly by applying deep learning methods based on various architectures (CNN,
LSTM, BERT) [
        <xref ref-type="bibr" rid="ref10 ref11 ref12">10, 11, 12</xref>
        ] and, less frequently, by the use of traditional machine learning methods
based on TF-IDF [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] or lexicon-based methods. Lexicon-based methods, despite their limitations
as demonstrated by experimental data, offer valuable insights into stylistic and lexical features,
making them useful tools for linguistic text analysis [
        <xref ref-type="bibr" rid="ref14">14, 15</xref>
        ]. Cross-lingual learning and translation
techniques have gained increasing prominence, providing effective solutions to the challenges of
text classification across diverse linguistic contexts [
        <xref ref-type="bibr" rid="ref15">16</xref>
        ]. Many studies have focused on English and
other languages with abundant resources, largely due to the availability of extensive datasets. In
particular, researchers have paid close attention to contextual embeddings, such as BERT and
fastText [
        <xref ref-type="bibr" rid="ref16">17</xref>
        ], which enable efficient handling of misspellings, rare words, and newly introduced
terms. Multilingual embeddings, like mBERT, further enhance the ability to process multilingual
offensive or toxic content [
        <xref ref-type="bibr" rid="ref17">18</xref>
        ].
      </p>
      <p>
        Recently, there have also been publications devoted to the detection of toxic language in
lowresource languages, including Ukrainian. This is primarily due to the release of multilingual large
language models (LLMs) and the development of new methods for creating datasets, including the
development of translated datasets and the generation of synthetic data. To date, there is no
publicly available expert-annotated toxic text dataset. The only existing corpus of this kind is the
Ukrainian tweets corpus [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ], which can be filtered by toxic keywords (Ukrainian obscene lexicon)
provided by the author. Thus, researchers are yet to find a solution to this issue. In their
publication [
        <xref ref-type="bibr" rid="ref18">19</xref>
        ], the authors describe three approaches to creating a corpus of Ukrainian-language
toxic texts with binary markup (toxic and non-toxic). (i) translation from English; (ii) toxic samples
filtering by toxic keywords; (iii) crowdsourcing data annotation for phrases containing five to
twenty words. Furthermore, researchers delineate and contrast three methodologies for identifying
toxicity: Prompting of LLMs, Cross-lingual transfer approach and Fine-tuning of LLMs on different
types of data. The results are somewhat controversial, as each approach demonstrates efficacy on a
different dataset. In [
        <xref ref-type="bibr" rid="ref19">20</xref>
        ], the authors present a bullying detection model for Ukrainian language. In
order to construct the model, the researchers created a dataset by means of machine translation
from English to Ukrainian. The authors assess the efficacy of the zero-shot technique and evaluate
the performance of contemporary multilingual models and embeddings (mBERT, XLM-R, LASER,
MUSE). As a result, the final detection model exhibits promising metrics. The authors conclude
that, in the context of low-resource languages, the classification accuracy of a given model tends to
increase in proportion to the number of samples used, regardless of their origin.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. A lexicon-based method for toxicity detection of Ukrainianlanguage texts</title>
      <p>The lexicon-based method is employed to determine the toxicity of a text. This entails identifying
toxic words within the text in accordance with pre-compiled dictionaries of toxic vocabulary. The
text is then evaluated on a toxicity scale according to the number of instances of toxic vocabulary
identified. This method is employed in conjunction with automatic morphological and syntactic
analysis, namely automatic rules-based text analysis. For this reason, the dictionary-based toxicity
assessment may also be referred to as the dictionary and lexicon-based method or the rules-based
method. The accuracy and completeness of the toxicity determination using this method are
contingent upon the scope and quality of the compiled lexicon, as well as the quality of the
lemmatization procedure for the text. The scope and semantics of the words in the toxic dictionary
are contingent upon the style, genre, and subject matter of the texts for which the analysis system
is being constructed.</p>
      <p>The efficacy of the dictionary-based approach is contingent upon its capacity to provide a
comprehensive linguistic analysis of text toxicity, culminating in the formulation of an expert
opinion. This entails the identification of a list of lexical toxic units that characterize the text in
question. In contrast, the considered machine learning method is limited to binary classification of
text toxicity based on the features and degree of toxicity. It is worth noting that the lexicon-based
method can be used to assess the toxicity of texts of varying lengths, even in the absence of toxic
datasets. One disadvantage of the method based on dictionaries and rules is the limited nature of
the lexicon. It is inherently incomplete because communication generates new means of expressing
toxicity and new toxic discourses that transform neutral vocabulary into the vocabulary of
destructive influence. This in turn requires the ongoing addition of new terms to the lexicon.</p>
      <sec id="sec-3-1">
        <title>3.1. Lexicographic lists of toxic words</title>
        <p>In the present study, our task was to compile lexical lists of toxic lexical means (including idiomatic
expressions) regardless of the topic of toxic texts in Ukrainian-language online media discourse.</p>
        <p>
          A toxic text is typically defined as an offensive comment or publication that exhibits one or
more of the following characteristics: harassment, threat, obscenity, cyberbullying, trolling,
indignation, and identity-based hate text [
          <xref ref-type="bibr" rid="ref10 ref11 ref12">10, 11, 12</xref>
          ]. The listed features are distinctly aggressive
pragmatic practices of the sender in the communicative process. However, in our opinion, such a
definition of a toxic text does not take into account the fact that a toxic text is the result of toxic
communication, which may not contain an aggressive pragmatic stance but may include
information about events, facts, phenomena that negatively affect the psychological and emotional
state of both the sender and the recipient, or express emotions and emotional evaluations that
reflect the psychological instability of the participants of the communication. Such texts have
become a feature of wartime media communication, containing words such as:
(occupier), (russia), etc.,
those that serve the function of destructive psychological influence;
(angry), etc., those that express negative evaluations.
Therefore, in our study, the concept of "toxic text" is interpreted in a somewhat broader manner: A
toxic text is defined as a text that not only contains indications of aggressive communication
(harassment, threats, obscenities, cyberbullying, trolling, outrage, and identity-based hate speech)
but also verbalizes negative facts, emotions, and assessments. These destructive
emotiongenerating words cause the recipient to experience anxiety, fear, confusion, shame, guilt,
oppression, and control. Within the concept of toxicity, we also differentiate between hate speech
as generally aggressive communication, and identity-based hate speech texts that display
aggression through discrimination based on a person's identity, such as nationality, race, skin
color, origin, gender, health status, sexual orientation, religiosity, or other features.
        </p>
        <p>
          Accordingly, the lexical lists of toxic vocabulary were compiled with the notion of this broader
interpretation of the concept of a "toxic text" based on the following data: 1) A textual sample of
approximately two million words [
          <xref ref-type="bibr" rid="ref20">21</xref>
          ], comprising texts from blogs, news sites, online publications,
comments to online publications from social networks, and so forth; 2) a database of semantic taxa
[
          <xref ref-type="bibr" rid="ref5">5</xref>
          ], compiled from Ukrainian journalistic texts totaling 40,000 words; 3) a tonality dictionary of
Ukrainian vocabulary compiled by O. Tolochko [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ]. The compilation of these lists was carried out
automatically using specially developed software for data search and import.
        </p>
      </sec>
      <sec id="sec-3-2">
        <title>3.1.1. The "Dictionary of Emotionogens"</title>
        <p>Lexicographic List No. 1, the "Dictionary of Emotionogens," contains over 5,000 lexemes that
verbalize negative facts, emotions, and assessments, causing the recipient to experience anxiety,
fear, confusion, shame, guilt, oppression, and control. The dictionary includes words as
independent parts of speech with a negative tonality (rated "-2" or "very negative"), namely nouns
(e.g., (debt), (ignore),
), adjectives (e.g., (difficult),
(nuclear)), adverbs
(e.g.,</p>
        <p>), verbs (e.g.,
subvert), (to ignore), (to nullify),
scam)) and adjectives formed from verbs (e.g.,
(knocked),</p>
      </sec>
      <sec id="sec-3-3">
        <title>3.1.2. The "Hate Speech Dictionary"</title>
        <p>(torn),
(to prohibit),
(to risk),
(squandered)).</p>
        <p>The lexical list, designated No. 2, "Hate Speech Dictionary," contains 3,000 lexemes that verbalize
aggressive communication, including harassment, threats, obscenities, cyberbullying, trolling,
(to
(to
(lesbo),
(scammer),
outrage, and identity-based hate text. The list covers the following lexical groups: negative names
for people 1,620 (e.g., (scammer),
(fatty), (torturer),
), obscene vocabulary</p>
        <p>613 (e.g.,
(fucker),
(to fuck up)) and abusive vocabulary and vulgarisms 787 (e.g.,
shitstorm), (bastard)). Each item in the list is marked with
semantic characteristics according to the developed classifications. For example, in the dictionary
"Negative Names for People," the lexemes are grouped into 18 categories that correspond to key
semantic features: negative designations of a person based on age ( ), sexual
orientation ( (queer), (lesbian), ), gender
( ), appearance ( (beanpole), (dystrophic),
(fatty)), antisocial behavior ( (swindler), (swindler), (bandit),
(clown), (gangster), (kingpin), (traitor)), social activity / passivity
( - (defector),
(indifferent)), nationality (
(defector),
(banderite),
(negro)), social status ( (bastard),</p>
        <p>), profession and financial status (
(bankrupt), ), political affiliation ( (nazi),
(leftist)), place of residence ( (pedorussia), (hillbilly)), religion
( (atheist), (sectarian)), intellectual abilities (
(idiot), (hack), ), diseases (
(down)), and comparison to plants or animals ( (amoeba), (sheep), (otter),
(lamb), (pig)) etc. Each word of hate speech received has been
annotated as belonging to one of the following categories: s sexism, r racism, e ageism, etc.</p>
      </sec>
      <sec id="sec-3-4">
        <title>3.1.3. Dictionary of Toxic Compounds</title>
        <p>Lexicographic List No. 3, "Toxic Compounds," encompasses 1,500 stable phrases that idiomatically
reflect toxic sentiment. The items in the list are classified by 26 semantic features, with each
combination assigned a semantic label of toxicity (T), identity-based hate speech (IBYS), or
expressiveness (E), e.g., IBYS 4. Claims of inferiority, moral flaws ( (dirty gypsy),
) IBYS 7. Mention in a derogatory or insulting context, obscene
comments ( (lousy
(pigs have no word around here),
) IBYS 10.</p>
        <p>Threats of physical destruction or any violence (
(death to Jews))
)
Bullying ( (degenerate brat))
and malicious jokes aimed at eliciting an emotional response from readers, or trolling (
(supreme traitors),
etc.(
tonality (
(rotten essence),
(atrophied brain),
), etc.</p>
        <p>)</p>
        <sec id="sec-3-4-1">
          <title>Ph. Phraseologisms with a toxic ) 72</title>
        </sec>
      </sec>
      <sec id="sec-3-5">
        <title>3.2. Implementation within the TextAttributor 1.0 system</title>
        <p>The described lexicon-based method for toxicity detection in Ukrainian-language texts was
implemented
within the TextAttributor 1.0 system
as a rule-based
module. The general
characteristics of the rule-based toxicity detection module are as follows: the input is user-provided
text, while the output is: 1) a numerical value of the statistical index of text toxicity; 2) the absolute
frequencies of toxic words and phrases classified by semantic classes; 3) the text with toxic words
and phrases highlighted.</p>
      </sec>
      <sec id="sec-3-6">
        <title>3.2.1. Automatic Analysis Algorithm</title>
        <p>The linguistic statistical analysis of toxicity is performed for each text using the program code for
calling the analyzers in the following sequence of actions:
1.
2.
3.
4.</p>
        <p>The text is tokenized into separate sentences and words;
Morphological annotation of words;
Followed by a contextual analysis that refines the morphological annotation codes;
Lemmatization of words;</p>
        <p>Identification of toxic vocabulary in the analyzed text: comparison of words in the text
with the registry of the toxic lexicon database and calculation of the absolute frequencies of toxic
words and phrases;</p>
        <p>Calculation of the text toxicity index;</p>
        <p>Generation of a linguistic expert report: a statistical map of toxic vocabulary in addition to
the text where toxic words and phrases are highlighted.</p>
      </sec>
      <sec id="sec-3-7">
        <title>3.2.2. Toxic Vocabulary Database</title>
        <p>The rule-based module operates on the basis of a lexicographic database of toxic tabled vocabulary,
compiled from three lexicographic lists. The respective table (developed based on MS SQL Server)
contains 9,500 rows.</p>
        <p>Description of the data structure of the table:
[wid]
[did]
[wrd]
record identifier;
dictionary identifier (used for grouping phenomena by dictionary);
word form or lemma; if the column
then this column would be a lemma;
[mitka]</p>
        <p>sub-category label;
[updateDate]</p>
        <p>record update date;
[updatedBy] the user who updated the record;
[sdcatitemID]</p>
        <p>category identifier;
[wrd2]</p>
        <p>next words of the phrase after the first (if
phrase, if `sltype=1`;
[wrd3]</p>
        <p>the third lemma of the phrase, if
[sltype] an indicator that this line contains a phrase stored as text or as lemmas.</p>
      </sec>
      <sec id="sec-3-8">
        <title>3.2.3. Text toxicity index</title>
        <p>The text toxicity index (Itox) is calculated using the following formula:


= 10 ( + | |(</p>
        <p>+  ))⁄ ,
where: e
the number of emotionogen words (Lexicographic List No. 1), m
the number of
hate speech words (Lexicographic List No. 2), t
the number of toxic phrases (Lexicographic List
No. 3) in the text analyzed by the system; K</p>
        <p>an intensifier of aggressive toxic speech units (K = 2);
n
number of words in the analyzed text; 10
normalization factor for the statistical parameter.</p>
        <p>(1)</p>
        <p>The formula for the statistical parameter the text toxicity index takes into account the usage
frequency for various classes of vocabulary in the text, differentiated by the semantic features of
three lexicons: emotionogenic words (e), hate speech lexemes (m), toxic phraseological compounds
(t). To enhance the weight of toxic means of aggressive communication, a coefficient K (on a
fivepoint scale [-2, -1, 0, +1, +2] it corresponds to -2) has been introduced into the formula. This
coefficient intensifies the weight of hate speech words (Lexicographic List 2) and toxic
phraseological compounds (Lexicographic List 3) in determining the level of toxicity. The
procedure of multiplying by 10 has been included into the formula to increase the empirical value
of the toxicity index for normalizing the linguistic statistical parameters of the stylometric analysis
in the TextAttributor 1.0 system.</p>
      </sec>
      <sec id="sec-3-9">
        <title>3.2.4. Results of the Rule-Based Toxicity Detection Module</title>
        <p>Let us consider the results of toxic speech analysis using a sample text: «
form, rather than my Ukraine"), which consists of 52 sentences and 500 words. The toxicity index of
the analyzed text (Itox) has been assessed at 0.7, indicating a slightly higher level of toxicity
compared to the upper limit of the average toxicity value for the media style in the Ukrainian
language (Fig. 1).</p>
        <p>The system generates a statistical map of the linguistic expert report on text toxicity (Fig. 1)
which presents a list of word semantic classes identified by the system in the analyzed text,
according to the classification markers of the toxic vocabulary database: Semantic category
column 1; name of semantic feature column 2; number of words/phrases in the analyzed text
according to semantic features column 3. In particular, the statistical map of the text "In 2019, a
completely different country began to form, rather than my Ukraine" systematizes the following
linguistic data: emotionogens 18; negative names for a person based on intellectual ability 2;
vulgarisms 2; negative names for a person by comparison with mythical creatures 1; negative
names for a person based on health characteristics 1; negative names for a person (sarcasm,
idiomatic expression) 1; negative names for a person based on body parts or physiological
processes 1.</p>
        <p>The verbalization of statistics for identified toxic vocabulary in the text, presented in the
statistical map according to semantic categories, is visualized in a separate window using bold
black font marking specific lexical means (Fig. 2).</p>
        <p>By comparing the statistical map data with the text, one can systematize toxic means, for
example: emotionogens 18 (nouns: (horror), (victim),
(cruelty), (corruption), (propaganda),
; adjectives: (painful), (uncontrollable),
(doomed), (terrible); adverb: (unfortunately); verbs:</p>
        <p>); vulgarisms 2 ( (finished));
toxic phrase (to kiss ass); negative person descriptors based on various characteristics
( (idiot), ).</p>
        <p>The provided text also includes indication of other features: 1) phraseoligisms are highlighted in
blue font ( etc.); some words and
symbols not recognized by the system's automatic morphological analysis are highlighted in red
font ( ). The
system does not consider these units when calculating the text toxicity index. Testing the system's
performance demonstrates the need to refine the lexicographic database, especially regarding
enhance the text preprocessing and lemmatization procedure. We consider this one of the primary
prospective tasks.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Machine learning methods for toxicity detection of Ukrainianlanguage texts</title>
      <p>4.1. Data
For our experimental study of toxicity in media texts using machine learning methods, two
datasets containing short Internet texts from Ukrainian-language blogs, comments, articles, etc.,
have been prepared.</p>
      <p>Dataset 1. For the initial analysis of the issue, a trial corpus of media texts was formed,
consisting of 668 documents featuring hate speech and expert annotations (each text sample was
classified by an expert into a specific category neutrality or hate speech). 226 texts were
identified as toxic. The obtained dataset was randomly split into a training set (600 texts, 192 of
which are toxic, 94,905 word samples, 11,852 stems) and a test set (68 texts, 34 of which are toxic,
10,800 word samples).</p>
      <p>Dataset 2. The training set was augmented by incorporating annotated 11,387 text documents,
among which 2,155 are toxic. The test set remained unchanged.</p>
      <sec id="sec-4-1">
        <title>4.2. Linguistic Features</title>
        <p>The typical pipeline of Machine Learning methods involves preprocessing the text data, extracting
features, and then using a classifier to determine if the text is toxic. Feature extraction techniques
include:</p>
        <p>• Bag-of-Words (BoW) is a simple approach where text is represented as an unordered
collection of words, ignoring grammar and word order but preserving frequency.</p>
        <p>• Term Frequency-Inverse Document Frequency (TF-IDF) enhances the BoW model by
weighing word frequency relative to its importance across documents, thus reducing the weight of
commonly used words.</p>
        <p>• Pre-trained word embeddings like Word2Vec or GloVe can be used to represent words as
dense vectors, capturing semantic similarities.</p>
        <p>• N-grams capture short phrases or word sequences to better model context compared to
BoW, useful for identifying common toxic word patterns.</p>
        <p>• Other linguistic features may include sentiment scores, lexical resources like swear word
dictionaries, Part-of-Speech (POS) tags, and syntactic features.</p>
        <p>• In experimental research described in this paper we rely on BoW and word embeddings.</p>
        <p>In experimental research described in this paper we rely on BoW and word embeddings.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.3. Classical Machine Learning Algorithms</title>
        <p>Classical, or primitive, Machine Learning Approaches use linguistic features and classical
algorithms like SVM, logistic regression and Naive Bayes. They require significant domain
knowledge and are limited in handling complex language patterns.</p>
        <p>Naive Bayes: Suitable for text classification due to the assumption of word independence, which
often works well despite its simplicity.</p>
        <p>Logistic Regression: A widely used linear model for binary classification.</p>
        <p>Support Vector Machines (SVM): Works well with high-dimensional data like text, especially
when combined with kernel methods to separate non-linear toxic and non-toxic classes.</p>
        <p>Random Forest and Decision Trees: Tree-based models are sometimes used to capture
nonlinear relationships, but they tend to struggle with sparse data in large text corpora.</p>
        <p>In this paper we describe experiments exploiting the Naive Bayes approach.</p>
      </sec>
      <sec id="sec-4-3">
        <title>4.4. Deep Learning Methods</title>
        <p>Deep Learning Approaches leverage neural networks, RNNs, CNNs, LSTMs, and especially
transformers like BERT for sophisticated contextual understanding of text. These models can
automatically learn features but require substantial computational resources and large datasets.</p>
        <p>Hybrid Approaches combine the strengths of both classical and deep learning methods for
better performance and robustness.</p>
        <p>Challenges across both methods include handling nuanced language, interpretability, data
imbalance, and avoiding bias.</p>
        <p>
          In this paper our choice is fastText [
          <xref ref-type="bibr" rid="ref21">22</xref>
          ], an efficient and widely-used method for word
representation and text classification, which can be used as a feature extraction technique in text
toxicity detection.
        </p>
        <p>FastText's main advantages are its relative simplicity, speed, and ability to handle large
vocabularies and datasets. It is particularly efficient because it:
• uses rather shallow neural network, which is much faster than deep models,
• embeds subword (character n-grams) information, allowing better generalization,
• utilizes hierarchical softmax for computational efficiency during classification, especially
for large-scale problems.</p>
        <p>The subword presentation is important since the Ukrainian language is highly inflective, and
plenty of unseen words as well as words with spelling errors must be covered. The effectiveness of
the selected approach is also determined by the technical conditions of system operation and
available textual resources for model training.</p>
      </sec>
      <sec id="sec-4-4">
        <title>4.5. LLM Prompting</title>
        <p>Text toxicity detection using Large Language Models (LLMs), such as GPT-3 or GPT-4, is an
emerging area in Natural Language Processing (NLP) that leverages advanced capabilities of LLMs
for identifying harmful, abusive, or toxic language in text. Here are the methods and strategies for
using LLM prompting to detect toxicity:
•
•
•
•
•
•
•</p>
        <p>Direct Toxicity Classification Prompting
Chain-of-Thought Prompting (Step-by-Step Reasoning)
Few-Shot Learning (Providing Examples)
Zero-Shot Prompting with Contextual Framing
Prompt Tuning for Toxicity Detection
Debiasing and Calibration Prompting</p>
        <p>Toxicity Detection as Part of a Larger System (Hybrid Models)</p>
        <p>
          In this work we consider one of the most straightforward methods involves directly prompting
the LLM to classify text as "toxic" or "non-toxic" exploiting Llama v. 3 [
          <xref ref-type="bibr" rid="ref22">23</xref>
          ]. This was tested by
providing the LLM with a clear instruction such as:
        </p>
        <p>Classify the following sentence as 'toxic' or 'non-toxic':
"You are an idiot and nobody likes you."</p>
        <p>The LLM is provided with text snippets, and based on its training and understanding of
language, it identifies the toxic content. Implementation of this approach is quick and simple, no
fine-tuning required. Another advantage is that LLMs can identify nuanced and context-dependent
toxicity that may escape simpler classifiers. On the other hand, (a) LLMs may not always be
consistent or reliable without additional constraints or examples, (b) false positives or negatives
could occur due to ambiguity or complexity in language, (c) huge amounts of computation
resources are required, including a powerful GPU with 24GB of RAM.</p>
      </sec>
      <sec id="sec-4-5">
        <title>4.6. Results</title>
        <p>Evaluating models for toxicity detection requires a careful choice of metrics because of the class
imbalance and Precision, Recall, and F1-Score meets these requirements: precision helps in
minimizing false positives (important to avoid over-censoring), while recall ensures detection of
most toxic content and F1 generalizes these two metrics.</p>
        <p>• Classical techniques: The following results were obtained from the testing of Dataset 1 and
2: F1 71 %; precision 60 %; recall 85 %.</p>
        <p>• Deep Learning: The best result according to the generalized metric (F1 = 79.4%) was
obtained for the following hyperparameter values: the dimension of the space of the vector
representation of words 56, the initial learning rate 0.15, the number of epochs 500, the
lexical context bigrams, the length of subwords from 2 to 5, the decision-making threshold
0.4. At the same time, for a sensitivity of 85%, an accuracy of about 70% is achieved.
• LLM prompting: The following results were obtained on the test sample: F1
precision 74.3 %; recall 76.5 %.</p>
      </sec>
      <sec id="sec-4-6">
        <title>4.7. Integration with the TextAttributor 1.0 system</title>
        <p>The experimental system for determining the toxicity of media texts based on ML is implemented
in a "client-server" architecture using a REST interface. A basic web user interface has been
developed, allowing users to input text and receive a response from the system on whether the text
is toxic (__label__target) with a confidence score ranging from 0 to 1. Figure 3 shows an example of
using the basic web interface to assess the toxicity of a given media text. In turn, using the REST
API, TextAttributor acts as a client, sending the user-entered text to the server, receiving a
response, and visualizing the result along with other parameters calculated for the given text.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Experiments and discussions</title>
      <p>Let us consider the experimental comparison of the toxicity indices of a single text, determined by
the two methods considered.</p>
      <sec id="sec-5-1">
        <title>5.1. Systematization of experimental data and the gradation of the toxicity index</title>
        <p>The experimental study aims to compare the results of the system's performance using two
methods rules-based and machine learning to determine the effectiveness of these methods in
identifying toxic texts. To achieve this goal, a sample of 520 texts was compiled and analyzed using
the TextAttributor 1.0 system. The sample included texts of varying lengths (from 33 words to
46,000 words), encompassing different themes, styles, and genres. Each text was evaluated for
toxicity using two indices: ToxR is the toxicity index based on the lexicon and rule-based method,
while ToxML is the toxicity index as determined by the machine learning method.</p>
        <p>Using the two methods in the TextAttributor 1.0 system, toxicity grading occurs on two scales:
ToxR is graded on a scale from 0 to 5.78, while ToxML, based on machine learning, ranges from 0
to 1.00. ToxML is limited to a range of 0 1.00, with a decision threshold for text toxicity set at 0.4
(0.35). (neutral text) 0 &lt;
lowdetermined experimentally based on the best F1 score for the trained model. In contrast, ToxR has
no upper limit or decision-making threshold value for the system. Establishing toxicity using this
method involves the identification of lexical markers of text toxicity, their statistical evaluation
the text toxicity index while the decision on toxicity is left to the linguist.</p>
        <p>During the linguistic analysis of the toxicity of texts examined by the system, the average ToxR
value for the media style of the Ukrainian language was empirically measured at 0.4 (0.35). This
value will be considered as the decision threshold on a text's toxicity, despite the fact that this
numerical value results from statistical analysis of entirely different features than those used in the
calculation of ToxML. Empirical numerical values of toxicity indices, determined by two methods
for each text in the sample, were ranked in descending order of numerical values: from highest to
lowest. Ranking the toxicity indices, calculated by the two methods, allows for the distribution of
texts according to toxicity levels:
•
•</p>
        <p>ToxML: toxic texts with an index of 0.4+ constitute 52.7% of the sample (276 texts).</p>
        <p>ToxR: toxic texts with an index of 0.4+ constitute 58.8% of the sample (306 texts).</p>
        <p>The remaining texts in the sample exhibit either a low degree of toxicity (0 &lt; 0.4) or are devoid
of toxic characteristics (neutral text = 0):</p>
        <p>• ToxML: low-toxicity texts with an index below 0.4 constitute 37.7% of the sample (196
texts); neutral texts with a toxicity level of 0 constitute 9.6% (50 texts).</p>
        <p>• ToxR: low-toxicity texts with an index below 0.4 constitute 37.5 % of the sample (195 texts);
neutral texts with a toxicity level of 0 constitute 3.7 % (19 texts).</p>
        <p>The percentage ratio of toxic and low-toxicity texts, as determined by the two methods,
demonstrates close results; however, the numerical toxicity indicators determined by the two
methods for a single text can be drastically different. Ranking texts by descending numerical values
of ToxML while preserving the numerical values of ToxR for these texts indicates that:
1. 276 texts are toxic by ToxML; however, 84 of these texts exhibit low toxicity (ToxR &lt; 0.4)
or are non-toxic (ToxR = 0) according to the ToxR method, meaning only 192 texts are assessed as
toxic by both methods.</p>
        <p>2. 244 texts are low-toxicity or neutral (ToxML &lt; 0.4), but 107 of these texts exhibit high
the system as low-toxicity/non-toxic by both methods.</p>
        <p>Thus, it can be stated that 329 out of 520 texts were classified as toxic/low-toxicity/emotionally
neutral by both methods. Thus, we are presented with a question: Is there a statistical relationship
between the two variables empirical numerical values of the toxicity indices determined by the
two methods?</p>
      </sec>
      <sec id="sec-5-2">
        <title>5.2. Pearson correlation between toxicity indices determined by two methods</title>
        <p>Comparing the two sets of numerical data ToxML and ToxR toxicity indices we have
formulated a task to determine whether there is a statistical dependency in the changes of ToxML
and ToxR toxicity levels on a sample of 520 texts, i.e., to ascertain the degree of correlation
between the two data sets: the ranked list of ToxML and corresponding numerical values of ToxR.
The correlation between the two data sets was measured using three methods: linear Pearson
correlation (0.2068); Spearman rank correlation (0.2513); Kendall rank correlation (0.1753).</p>
        <p>Let us consider the value of Pearson's linear correlation coefficient, which is most frequently
used in humanities and social studies to analyze correlation rather than causal dependency. The
correlation coefficient r = 0.2068 is low and indicates a weak positive correlation. However, with an
error p = 0.001, considering the degrees of freedom f = 600, the critical correlation coefficient value
is r = 0.134. The empirical value of Pearson's coefficient, r = 0.2068, is higher than the critical value
and demonstrates high significance for a sample size of 520 texts. This implies that, although the
degree of correlation is weak, the indicator has high reliability.</p>
        <p>Let us consider a scatter plot for Pearson's coefficient of variation (0.2068), which indicates a
weak measure of joint variability between two random variables two toxicity indices, ToxML and
ToxR, determined by two methods for a single text (Fig. 4). For the convenience of graphical
modeling of correlation, the empirical values of ToxR were converted to a scale from 0 to 1 using
the min- norm = (Xi - Xmin) / (Xmax Xmin).</p>
        <p>X-axis: numerical values of toxicity levels determined by the machine learning method (ToxML
independent variables); Y-axis, the numerical values of the toxicity index determined by the
rulebased method (ToxR dependent variables). Each point on the diagram represents a text, with its
coordinates being two toxicity indices. The red line is a regression line modeling the relationship
between the two variables. Given a weak positive Pearson coefficient, the regression line has a low
slope. This indicates a vague trend of increasing Y values with the increase in X values.
Considering the decision threshold of 0.4 for determining text toxicity/low toxicity for ToxML and
0.07 (after min-max normalization) for ToxR, four zones were identified on the graph: 1) Texts
toxic by ToxR, but low-toxicity or neutral according to ToxML dark gray zone; 2) Texts toxic
according to both methods light gray zone; 3) Texts considered low-toxicity or neutral according
to both methods light gray zone; 4) Texts toxic by ToxML, but low-toxicity or neutral by ToxR
dark gray zone. The correlation of variables (ToxR and ToxML) for texts in zone 1 and zone 4
consequently affects the coefficient of variation, as an inversely proportional correlation is
observed in these zones. Low X values correspond to high Y values (Zone 1), whereas high X values
correspond to low Y values (Zone 4). We consider the tests of these areas to be contentious with
respect to toxicity indicators.</p>
        <p>The results of automatic toxicity analysis also demonstrate similar statistical data by the two
methods in determining the toxicity/low-toxicity/neutrality of texts, which, according to the scatter
plot, are located in zones 2 and 3. To determine the degree of correlation between the statistical
data obtained using the two methods for these texts, Pearson's correlation coefficient was
calculated for a sample of 329 texts, of which 192 are toxic, 133 are low-toxicity, and 4 are neutral
texts. The correlation coefficient r = 0.5344 is relatively high and indicates a moderate positive
correlation. Moreover, the average degree of correlation has high reliability, as with p = 0.001 and
considering the degrees of freedom f = 350, the critical correlation coefficient value is r = 0.175. The
empirical value of Pearson's coefficient, r = 0.5344, is higher than the critical value and
demonstrates high significance for a sample size of 329 texts.</p>
        <p>A scatter plot was also constructed for Pearson's coefficient of variation (0.5344) (Fig. 5).</p>
        <p>The regression line, which models the average positive correlation coefficient on the diagram,
has a significantly greater slope compared to the diagram in Fig. 4, which indicates a moderate
measure of joint variability between the two random variables ToxR and ToxML. This indicates a
clear trend of increasing Y values with the increase in X values.</p>
        <p>In Figure 5, four zones were identified using the same principle: 1) Texts toxic by ToxR, but
lowtoxicity or neutral according to ToxML - dark gray zone; 2) Texts toxic according to both methods
light gray zone; 3) Texts considered low-toxicity or neutral according to both methods - light gray
zone; 4) Texts toxic by ToxML, but low-toxicity or neutral by ToxR - dark gray zone. The
correlation of variables (ToxR and ToxML) shows that texts rarely fall into zones 1 and 4; instead,
the points in zone 2 clearly represent low-toxicity and neutral texts, while zone 4 contains points
representing texts with high toxicity. With an average positive correlation coefficient (Fig. 5), the
dispersion of points is significantly lower than with a weak positive coefficient (Fig. 4), yet in both
cases, "outliers" from the model's range of inference are observed. These are points of texts that
deviate from the general trend of variable correlation and are typically characterized by high levels
of toxicity according to one or both methods.</p>
        <p>The diagram also clearly shows the formation of three clusters: 1) Cluster of points in the range
of 0 0.3 (on the X-axis) represents neutral texts and texts with low toxicity; 2) Cluster of points in
the range of 0.8 1.0 (on the X-axis) represents texts with high toxicity; 3) Dispersion of points in
the range of 0.4 0.8 represents texts of medium toxicity. In the range of 0.3 0.4, there are 5 points
texts whose degree of toxicity is difficult to interpret because they fall within the decision
threshold range. However, 2 points are closer to cluster 1 (neutral texts and low-toxicity texts),
while 3 points are closer to cluster 3 (medium toxicity texts).</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>6. Conclusions and future work</title>
      <p>promising outcomes, particularly in the automatic identification of toxic content in Ukrainian texts.
The system combines a rule-based and a machine learning approach, both of which provide
valuable, yet distinct, insights into text toxicity. The rule-based module excels in providing a
detailed linguistic analysis of toxic vocabulary based on a lexicographic database, making it
suitable for in-depth expert analysis. On the other hand, the machine learning module provides a
scalable solution for handling large volumes of text, offering an efficient and automated method of
classifying toxic content.</p>
      <p>However, a moderate correlation between the two methods, as demonstrated by the Pearson
correlation coefficient, reveals some discrepancies in toxicity assessment. These differences
underline the need for further refinement in both modules to improve the overall system accuracy
and reliability.</p>
      <p>Key tasks for further improvement include enhancing the machine learning model to analyze
full-length texts rather than truncated segments and expanding its training dataset to increase
accuracy for longer texts. Additionally, testing more advanced models such as those from the BERT
family could further improve classification performance. On the rule-based side, recalibration of
the toxicity index formula is required to address limitations in how text size and single occurrences
of aggressive language impact the final score.</p>
      <p>One of the most promising directions for future work is the integration of rule-based and
machine learning methods into a hybrid model. Such a model would leverage the strengths of both
approaches, applying rule-based analysis for in-depth, context-sensitive interpretation and
machine learning for efficient large-scale classification. This hybrid approach could also enhance
the system's ability to detect more subtle forms of toxicity, such as covert hate speech or
contextdependent insults.</p>
      <p>Further development of the user interface to display toxicity results more intuitively would
-technical users. Visual tools such as heatmaps of toxic word
distributions, graphs showing the progression of toxicity throughout a text, and interactive
dashboards could make the system more accessible for use in various domains.</p>
    </sec>
    <sec id="sec-7">
      <title>Declaration on Generative AI</title>
      <sec id="sec-7-1">
        <title>The authors have not employed any Generative AI tools.</title>
        <p>183.
http://iti.fit.univ.kiev.ua/wp</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1] Dictionary of media literacy,
          <year>2022</year>
          . URL: https://filter.mkip.gov.ua/wp-content/uploads/ 2022/10/slovnyk.pdf [in Ukrainian].
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>V.M.</given-names>
            <surname>Petryk</surname>
          </string-name>
          ,
          <string-name>
            <surname>M.M. Prysiazhniuk</surname>
            ,
            <given-names>L.F.</given-names>
          </string-name>
          <string-name>
            <surname>Kompantseva</surname>
          </string-name>
          , Ye.D.
          <string-name>
            <surname>Skulysh</surname>
            ,
            <given-names>O.D.</given-names>
          </string-name>
          <string-name>
            <surname>Boiko</surname>
            ,
            <given-names>V.V.</given-names>
          </string-name>
          <string-name>
            <surname>Ostroukhov</surname>
          </string-name>
          .
          <article-title>Suggestive technologies of manipulative influence: study guide</article-title>
          .
          <source>Kyiv</source>
          ,
          <year>2011</year>
          [in Ukrainian].
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <surname>Recommendation</surname>
            <given-names>CM</given-names>
          </string-name>
          / Rec (
          <year>2022</year>
          )
          <article-title>16 of the Committee of Ministers to member States on combating hate speech</article-title>
          . URL: https://www.coe.int/en/web/combating- hatespeech/recommendation-on
          <article-title>-combating-hate-speech.</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          <source>[4] TextAttributor 1</source>
          .0,
          <year>2024</year>
          . URL: http://ta.mova.info
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>N.</given-names>
            <surname>Darchuk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Zuban</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Lanhenbakh</surname>
          </string-name>
          , Ya. Khodakivska.
          <article-title>AGAT-semantics: semantic markup of the Ukrainian language corpus</article-title>
          .
          <source>Ukrainian linguistics</source>
          ,
          <volume>1</volume>
          (
          <issue>46</issue>
          ),
          <year>2016</year>
          ,
          <volume>92</volume>
          102. https://doi.org/10.17721/um/46 [in Ukrainian].
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>N.</given-names>
            <surname>Darchuk</surname>
          </string-name>
          .
          <article-title>Linguistic approach for development of computer-based sentiment analysis in the Ukrainian language</article-title>
          .
          <source>Science and education a new dimension</source>
          ,
          <volume>189</volume>
          , (
          <year>2019</year>
          ),
          <fpage>10</fpage>
          <lpage>13</lpage>
          [in Ukrainian].
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>O.</given-names>
            <surname>Zuban</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Bilous</surname>
          </string-name>
          .
          <article-title>Automatic determination of lexical tone of Ukrainian-language text content/uploads/ITI-2018</article-title>
          .pdf [in Ukrainian].
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>O.</given-names>
            <surname>Tolochko</surname>
          </string-name>
          .
          <article-title>Tonal dictionary of the Ukrainian language</article-title>
          ,
          <year>2023</year>
          . URL: https://github.com/Oksana504/sentimentdictionary-uk.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>K.</given-names>
            <surname>Bobrovnyk</surname>
          </string-name>
          .
          <article-title>Automated building and analysis of ukrainian twitter corpus for toxic text detection</article-title>
          .
          <source>In COLINS 2019. Volume II: Workshop</source>
          ,
          <year>2019</year>
          , pp.
          <fpage>55</fpage>
          -
          <lpage>56</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>F.</given-names>
            <surname>Alkomah</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Ma</surname>
          </string-name>
          .
          <article-title>A Literature Review of Textual Hate Speech Detection Methods</article-title>
          and Datasets,
          <source>Information 13.6</source>
          (
          <year>2022</year>
          ). doi:
          <volume>10</volume>
          .3390/info13060273.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>Jahan</surname>
            ,
            <given-names>M. S.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Oussalah</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <article-title>A systematic review of hate speech automatic detection using natural language processing</article-title>
          . Neurocomputing,
          <volume>546</volume>
          (
          <year>2023</year>
          ),
          <volume>126232</volume>
          . doi:
          <volume>10</volume>
          .1016/j.neucom.
          <year>2023</year>
          .
          <volume>126232</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>A.</given-names>
            <surname>Bonetti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Martínez-Sober</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. C.</given-names>
            <surname>Torres</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Vega</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Pellerin</surname>
          </string-name>
          , and J.
          <string-name>
            <surname>Vila-Francés</surname>
          </string-name>
          ,
          <article-title>Comparison between Machine Learning and Deep Learning Approaches for the Detection of Toxic Comments on Social Networks</article-title>
          ,
          <source>Applied Sciences 13.10</source>
          (
          <year>2023</year>
          ). doi:
          <volume>10</volume>
          .3390/app13106038.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>F.</given-names>
            <surname>Rangel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Sarracén</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Chulvi</surname>
          </string-name>
          , E. Fersini,
          <string-name>
            <given-names>P.</given-names>
            <surname>Rosso</surname>
          </string-name>
          .
          <article-title>Profiling Hate Speech Spreaders on Twitter Task at PAN 2021</article-title>
          .
          <source>In Proceedings of the CLEF 2021 Conference and Labs of the Evaluation Forum</source>
          ,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>S.</given-names>
            <surname>Frenda</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Ghanem</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Montes-Y-Gómez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Rosso</surname>
          </string-name>
          .
          <article-title>Online hate speech against women: Automatic identification of misogyny and sexism on twitter</article-title>
          .
          <source>J. Intell. Fuzzy Syst</source>
          .
          <volume>36</volume>
          (
          <year>2019</year>
          ),
          <fpage>4743</fpage>
          <lpage>4752</lpage>
          . doi:
          <volume>10</volume>
          .3233/JIFS-179023.
          <article-title>processing in the detection of degrees of toxicity in online discussions</article-title>
          .
          <source>Sensors</source>
          ,
          <volume>22</volume>
          (
          <issue>17</issue>
          ) (
          <year>2022</year>
          ),
          <volume>6468</volume>
          . doi:
          <volume>10</volume>
          .3390/s22176468.
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>F.-Z.</given-names>
            <surname>El-Alami</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Ouatik El Alaoui</surname>
          </string-name>
          , and
          <string-name>
            <given-names>N.</given-names>
            <surname>En-Nahnahi</surname>
          </string-name>
          .
          <article-title>A multilingual offensive language detection method based on transfer learning from transformer fine-tuning model</article-title>
          .
          <source>J. King Saud Univ. Comput. Inf. Sci.</source>
          ,
          <volume>34</volume>
          (8 Part
          <string-name>
            <surname>B)</surname>
          </string-name>
          (
          <year>2022</year>
          ),
          <fpage>6048</fpage>
          <lpage>6056</lpage>
          . doi:
          <volume>10</volume>
          .1016/j.jksuci.
          <year>2021</year>
          .
          <volume>07</volume>
          .013.
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [17]
          <string-name>
            <surname>A. G. D'sa</surname>
            ,, I. Illina,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Fohr</surname>
          </string-name>
          .
          <article-title>Bert and fasttext embeddings for automatic detection of toxic speech</article-title>
          .
          <source>In: 2020 International Multi-5. doi: 10.1109/OCTA49274</source>
          .
          <year>2020</year>
          .9151853
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>M. A. H.</given-names>
            <surname>Wadud</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. F.</given-names>
            <surname>Mridha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Shin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Nur.</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. K.</given-names>
            <surname>Saha.</surname>
          </string-name>
          Deep-BERT:
          <article-title>Transfer Learning for Classifying Multilingual Offensive Texts on Social Media</article-title>
          .
          <source>Computer Systems Science &amp; Engineering</source>
          ,
          <volume>44</volume>
          (
          <issue>2</issue>
          ) (
          <year>2022</year>
          ). doi:
          <volume>10</volume>
          .32604/csse.
          <year>2023</year>
          .
          <volume>027841</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>D.</given-names>
            <surname>Dementieva</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Khylenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Babakov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Groh</surname>
          </string-name>
          .
          <article-title>Toxicity Classification in Ukrainian</article-title>
          .
          <source>In Proceedings of the 8th Workshop on Online Abuse and Harms (WOAH</source>
          <year>2024</year>
          ),
          <year>2024</year>
          , pages
          <fpage>244</fpage>
          <lpage>255</lpage>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <year>2024</year>
          .woah-
          <volume>1</volume>
          .
          <fpage>19</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>V.</given-names>
            <surname>Oliinyk</surname>
          </string-name>
          and
          <string-name>
            <given-names>I.</given-names>
            <surname>Matviichuk</surname>
          </string-name>
          .
          <article-title>Low-resource text classification using cross-lingual models for bullying detection in the Ukrainian language</article-title>
          .
          <source>Adaptive systems of automatic control: interdepartmental scientific and technical collection</source>
          ,
          <volume>1</volume>
          (
          <issue>42</issue>
          ) (
          <year>2023</year>
          ). doi:
          <volume>10</volume>
          .20535/
          <fpage>1560</fpage>
          -
          <lpage>8956</lpage>
          .
          <fpage>42</fpage>
          .
          <year>2023</year>
          .279093
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [21]
          <article-title>Corpus of the Ukrainian language</article-title>
          ,
          <year>2003</year>
          2024.
          <article-title>Mova.info: linguistic portal</article-title>
          . URL: http://www.mova.info/corpus.aspx [in Ukrainian].
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>A.</given-names>
            <surname>Joulin</surname>
          </string-name>
          , E. Grave,
          <string-name>
            <given-names>P.</given-names>
            <surname>Bojanowski</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Douze</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Jégou</surname>
          </string-name>
          , T. Mikolov, FastText.zip:
          <article-title>Compressing text classification models</article-title>
          ,
          <source>arXiv preprint arXiv:1612.03651</source>
          ,
          <year>2016</year>
          . doi:
          <volume>10</volume>
          .48550/arXiv.1612.03651.
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [23]
          <string-name>
            <surname>Meta-Llama</surname>
          </string-name>
          .
          <article-title>Inference code for Llama models</article-title>
          ,
          <year>2024</year>
          . URL: https://github.com/metallama/llama/.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>