<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Abusive Speech Detection Method for Ukrainian Language Used Recurrent Neural Network</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Iurii</forename><surname>Krak</surname></persName>
							<email>yuri.krak@gmail.com</email>
							<affiliation key="aff0">
								<orgName type="institution">Taras Shevchenko National University of Kyiv</orgName>
								<address>
									<country key="UA">Ukraine</country>
								</address>
							</affiliation>
							<affiliation key="aff1">
								<orgName type="institution">Glushkov Institute of Cybernetics of NAS of Ukraine</orgName>
								<address>
									<settlement>Kyiv</settlement>
									<country key="UA">Ukraine</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Olha</forename><surname>Zalutska</surname></persName>
							<email>zalutska.olha@gmail.com</email>
							<affiliation key="aff2">
								<orgName type="institution">Khmelnytskyi National University</orgName>
								<address>
									<addrLine>11, Institutes str</addrLine>
									<postCode>29016</postCode>
									<settlement>Khmelnytskyi</settlement>
									<country>Ukraine Country</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Maryna</forename><surname>Molchanova</surname></persName>
							<email>m.o.molchanova@gmail.com</email>
							<affiliation key="aff2">
								<orgName type="institution">Khmelnytskyi National University</orgName>
								<address>
									<addrLine>11, Institutes str</addrLine>
									<postCode>29016</postCode>
									<settlement>Khmelnytskyi</settlement>
									<country>Ukraine Country</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Olexander</forename><surname>Mazurets</surname></persName>
							<affiliation key="aff2">
								<orgName type="institution">Khmelnytskyi National University</orgName>
								<address>
									<addrLine>11, Institutes str</addrLine>
									<postCode>29016</postCode>
									<settlement>Khmelnytskyi</settlement>
									<country>Ukraine Country</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Ruslan</forename><surname>Bahrii</surname></persName>
							<affiliation key="aff2">
								<orgName type="institution">Khmelnytskyi National University</orgName>
								<address>
									<addrLine>11, Institutes str</addrLine>
									<postCode>29016</postCode>
									<settlement>Khmelnytskyi</settlement>
									<country>Ukraine Country</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Olena</forename><surname>Sobko</surname></persName>
							<email>olenasobko.ua@gmail.com</email>
							<affiliation key="aff2">
								<orgName type="institution">Khmelnytskyi National University</orgName>
								<address>
									<addrLine>11, Institutes str</addrLine>
									<postCode>29016</postCode>
									<settlement>Khmelnytskyi</settlement>
									<country>Ukraine Country</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Olexander</forename><surname>Barmak</surname></persName>
							<email>alexander.barmak@gmail.com</email>
							<affiliation key="aff2">
								<orgName type="institution">Khmelnytskyi National University</orgName>
								<address>
									<addrLine>11, Institutes str</addrLine>
									<postCode>29016</postCode>
									<settlement>Khmelnytskyi</settlement>
									<country>Ukraine Country</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Abusive Speech Detection Method for Ukrainian Language Used Recurrent Neural Network</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">66B2D3C6DC2D00F224DD2291C65E1027</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T18:15+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>RNN, sentiment analysis, sentiment classification, social acceptance1 O. Barmak) 0000-0002-8043-0785 (I. Krak)</term>
					<term>0000-0003-1242-3548 (O. Zalutska)</term>
					<term>0000-0001-9810-936X (M. Molchanova)</term>
					<term>0000-0002-8900-0650 (O. Mazurets)</term>
					<term>0000-0001-5219-1185 (R. Bahrii)</term>
					<term>0000-0001-5371-5788 (O. Sobko)</term>
					<term>0000-0003-0739-9678 (O. Barmak)</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>The paper is devoted to the creation and approbation of a method for determining the level of social acceptability of textual Ukrainian-language content, which will be able to determine the level of detection of offensive speech for the Ukrainian language using a recurrent neural network based on the entered textual information. The method of detecting offensive speech involves the use of a combined approach based on the use of a dictionary of offensive words and a neural network approach to determine the sentiment tone of the message. The proposed method will allow to assess the level of social acceptability of Ukrainian-language Internet content for automated moderation of Internet content. Further research can be focused on applied use, which can be a useful tool for assessing the level of social acceptability of digital text content published on social networks and for preventing the spread of harmful or offensive information. It can also help improve the quality of communication and increase the level of social interaction in general.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>The terms and conditions of the vast majority of social networks and messengers clearly define content requirements and acceptable standards of user behavior. For violations of such rules and regulations, social networks usually block either content or accounts. The community standards of Facebook, Instagram, Twitter, YouTube, and other similar platforms prohibit the publication of terrorist content containing hate speech at various levels, prevent copyright infringement, coordinated inauthentic behavior, bullying, and the posting of pornographic materials <ref type="bibr" target="#b0">[1]</ref>. Social networks also fight against fake pages, and some platforms even delete inactive accounts. All social media platforms, including Facebook, Twitter, and YouTube, have made huge investments and developed policies to identify and moderate such harmful content <ref type="bibr" target="#b1">[2,</ref><ref type="bibr" target="#b2">3]</ref>.</p><p>The content that is considered socially acceptable can vary depending on various factors, such as cultural and moral norms, age groups, geographic location, and others. However, in general, socially acceptable content is content that does not violate the rules of a social network and does not cause disgust or outrage among a significant part of the audience <ref type="bibr" target="#b0">[1]</ref>.</p><p>Many published works are devoted to the detection of hate speech and offensive language in comments in English. However, according to the authors' research <ref type="bibr" target="#b3">[4]</ref>, there are no appropriate methods for automatically detecting obscenities and hate speech in Ukrainian, as the development of such methods is a difficult task for several reasons:  There are no labeled databases or Ukrainian-language corpora of comments or social media posts with offensive content.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head></head><p>Due to the peculiarities of word formation in the Ukrainian language, it is almost impossible to determine a definitive list of offensive words.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head></head><p>Users often invent new, explicitly obscene words by connecting two common stems with an infix or adding prefixes or suffixes.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head></head><p>The use of surzhik gives rise to a wide variety of swear words. The paper <ref type="bibr" target="#b4">[5]</ref> shows that a combination of machine learning and a lexicon-based approach can achieve higher accuracy than any type of sentiment analysis. The authors used a variety of sentiment analysis, machine learning methods, and dictionary-based sentiment analysis to test and compare the effectiveness of user behavior research.</p><p>The study <ref type="bibr" target="#b5">[6]</ref> was conducted on sentiment analysis on Twitter. Their proposed method is based on a dictionary and allows identifying sentiments about the AstraZeneca/Oxford, Moderna, and Pfizer/BioNTech COVID-19 vaccines for 4 months. A similar study was conducted <ref type="bibr" target="#b6">[7]</ref>, but the authors proposed to use TextBlob based on TF-IDF vectorization to assess sentiment. LinearSVC was chosen as the classification model, which resulted in an accuracy of 0.96752 for English-language tweets.</p><p>The study <ref type="bibr" target="#b7">[8]</ref> was aimed at identifying hate speech, which includes various forms of trolling, intimidation, harassment, and threats directed against specific individuals or groups of athletes. The experiments concern the detection of hate speech in the Serbian language. The proposed BiLSTM deep neural network <ref type="bibr" target="#b8">[9]</ref>, trained with different parameters, showed high accuracy in detecting hate speech in the sports sphere (96% and 97%) and a fairly low level of memorization.</p><p>The authors of paper <ref type="bibr" target="#b9">[10]</ref> implemented a special type of deep learning based on a recurrent neural network (RNN) <ref type="bibr" target="#b10">[11]</ref> called long-term memory (LSTM) for the automatic identification of hate and offensive content. The authors propose a language-agnostic solution for three Indo-European languages (English, German, and Hindi), the methodology does not use any pretrained model, which leads to a neutral language solution, and does not require the development of cumbersome features for the proposed model.</p><p>According to TIME <ref type="bibr" target="#b11">[12]</ref>, Facebook removed more than seven million cases of hate speech in the third quarter of 2019, a 59% increase over the previous quarter. More and more of these hateful statements (80%) are now detected not by humans, but automatically by artificial intelligence.</p><p>However, the algorithms Facebook currently uses to remove hate speech only work in certain languages. This means that it has become easier for Facebook to curb the spread of racial or religious hatred online in predominantly developed countries and communities dominated by global languages such as English, Spanish, and Chinese.</p><p>According to Time, Facebook automatically detects such statements in more than 40 languages around the world. In other languages, Facebook relies on its users and human moderators to control hate speech.</p><p>Contrary to the algorithms that Facebook says now automatically detect 80% of hate posts without requiring the user to report them first, these human moderators do not regularly scan the site for hate speech themselves. Instead, their job is to decide whether to remove posts that have already been reported by users <ref type="bibr" target="#b12">[13]</ref>.</p><p>Minority languages are the most affected by this inequality. This means that racially motivated slurs, calls for violence, and targeted insults can spread faster in developing countries than they currently do in the United States, Europe, and elsewhere.</p><p>Therefore, although virtual communication through social media platforms is an integral part of human life, there is a downside that comes in the form of harmful online content <ref type="bibr" target="#b13">[14]</ref>. Harmful content, whether it is fake news, rumors, hate speech, aggression, or cyberbullying, is a matter of serious concern to society <ref type="bibr" target="#b14">[15]</ref>. Such harmful content affects a person's mental health and also leads to losses that cannot be compensated for <ref type="bibr" target="#b15">[16]</ref>. Detection and moderation of such content are the primary tasks of information technology. The solution to the problem of determining the level of social acceptability of textual Ukrainian-language Internet content will allow automated moderation of Internet content <ref type="bibr" target="#b16">[17]</ref>.</p><p>Considering the analysis of the studies, we can distinguish approaches that allow automated detection of offensive language in textual Internet content for the Ukrainian language:  Sentiment analysis. This approach is used to detect and classify the emotional tone of the text, which helps to determine the overall mood of the message. For example, positive, negative.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head></head><p>Detection of abusive speech. Abusive speech detection tools are used to identify and classify texts that contain rudeness, insults, threats, etc.  Machine learning. The use of machine learning algorithms allows you to create models that can determine the social acceptability of a text based on training on large data sets. Such models can take into account a wide range of features, such as semantics, syntax, emotional connotation, and other factors that affect the detection of offensive speech <ref type="bibr" target="#b17">[18]</ref>. Offensive language is characterized by the presence of two aspects: 1. The presence of offensive words that define this content as offensive. 2. The presence of negative sentiment tone of content that actualizes offensive intentions. Accordingly, in order to detect offensive speech, it is necessary to evaluate the content under study for the presence of each of the manifestations.</p><p>The aim of research is to create and validate the method for automated determination of the level of social acceptability of textual Internet content based on a combined approach that includes sentiment analysis using RNNs and verification with a dictionary of offensive words.</p><p>The main contributions of this research are follows:  a method for detecting offensive language in textual Internet content for the Ukrainian language using a recurrent neural network has been developed;  the developed method allows detecting offensive speech with a given user threshold.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Methods and Materials</head><p>Given these limitations, there is a necessity to generate experimental data that will satisfy the research objectives.</p><p>The method for detecting offensive speech for the Ukrainian language using a recurrent neural network based on a combined approach that includes sentiment analysis and verification with a dictionary of offensive words is proposed. Its main stages of work are also described.</p><p>The proposed method has been validated in the format of an application implementation in Python, which clearly demonstrates the effectiveness of this approach. An efficiency study is also performed.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1.">Datasets</head><p>The «Ukrainian Twitter Corpus» (for RNN training) and the «AbusiveLanguageDataset» (for identifying offensive content) will be used as experimental data to implement the method of detecting offensive speech for the Ukrainian language using a recurrent neural network.</p><p>«Ukrainian Twitter Corpus» <ref type="bibr" target="#b18">[19]</ref> is a corpus of Ukrainian-language tweets collected using the Twitter API. The corpus contains more than 6 million tweets that were collected from December 2015 to May 2017.</p><p>Each tweet in the corpus is labeled as positive, negative, or neutral. Annotation was performed using machine learning algorithms and evaluated at the message level. In total, the proposed corpus contains about 3 million positive tweets, 1.7 million negative and 1.6 million neutral tweets. Of that total, 400 thousand tweets are in Ukrainian.</p><p>The corpus also contains additional data on the number of likes, retweets, and replies per tweet. The data is available for download on GitHub. As the data in the corpus is collected from the social network Twitter, it may contain informal vocabulary, abbreviations and other speech features that are typical for this source. However, given the topic of the study, it is appropriate to use this corpus to train an RNN that will perform the task of determining the semantic tone of a message.</p><p>«AbusiveLanguageDataset» <ref type="bibr" target="#b19">[20]</ref> is a dataset that contains Ukrainian-language comments with different levels of offensive content. This dataset consists of 5,000 comments, most of which are negative and contain foul language. The comments were collected from popular Ukrainian online resources, such as: «Ukrainska Pravda», «Tablo ID» and «TSN» <ref type="bibr" target="#b19">[20]</ref>. The dataset is presented in many languages, including Ukrainian. Each comment in the selected dataset was pre-evaluated by a human who determined whether it contains offensive content and the level of offensiveness (low, medium or high). In addition, for each comment, the identifier of the source from which the comment was received is indicated.</p><p>The dataset will be used as data for a numerical assessment of the level of social acceptability of the textual content of the posts.</p><p>The datasets were supplemented with data from previous studies, a dataset of tagged reviews from the hotline consisting of 7656 documents <ref type="bibr" target="#b20">[21]</ref>. The peculiarity of the dataset is that it contains russisms, swear words, and some reviews were presented in other languages. It is also supplemented by manually collected and labeled posts and comments from the social network Facebook in the amount of 500 units (250 positive and 250 negative).</p><p>Since even the tagged tweets and feedback from the hotline were in different languages, they were cleaned up with the help of the python language using the «langdetect library». The tweets and reviews were also filtered, and tweets consisting of less than 3 words were deleted. After imposing restrictions on tweets and Ukrainian-language reviews from the hotline, their number was as follows: the tweet set consisted of 2400 tweets (1200 positive and 1200 negative), the data set from the hotline consisted of 2000 positive and 2000 negative, the set of posts and comments from the social network Facebook consisted of 500 items (250 positive and 250 negative). The number of offensive words in the dictionary is 959 offensive words (Figure <ref type="figure" target="#fig_0">1</ref>). The described set of data will be used to implement a method for detecting offensive speech for the Ukrainian language, which will be able to determine the level of acceptability of usergenerated Internet content based on the entered textual information according to the established sensitivity threshold.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.">Components and stages of the method of detecting offensive speech for the Ukrainian language</head><p>In order to identify the level of offensive speech for Ukrainian-language Internet content, based on the analysis of publications, in particular, <ref type="bibr" target="#b21">[22]</ref> shows that studies that mainly relied on dictionary tools to extract sentiment from textual data and have a clear advantage in terms of interpretation, obviously lose in accuracy. Therefore, in order to ensure the reliability of the result in determining the level of social acceptability of textual Ukrainian-language Internet content, a neural network will be used to identify emotional content and a dictionary approach will be used to check for elements of unacceptable content.</p><p>The input data of the method are: a pre-trained RNN tone analysis model, textual Internet content for analysis, and a dictionary of key offensive expressions. The Internet content for analysis is characterized by some features <ref type="bibr" target="#b22">[23]</ref>, the main ones being:</p><p> short length (often has a limit on the number of characters, which causes the text to be short and requires a clear and concise expression of thoughts)  informative nature (usually, information is presented in an informal manner and may contain abbreviations, non-standard vocabulary, or surzhik);  use of multimedia (texts may include gifs, emojis, etc.);  dialogic nature (social networks have an extensive structure that facilitates interaction between users and the creation of dialogues). All this makes the processing of short online content specific, different from general methods for working with texts. A generalized diagram of the proposed method is illustrated in Figure <ref type="figure" target="#fig_1">2</ref>. Step 1 is a neural network assessment of the sentiment tone based on a pre-trained RNN model. The result of this step is a numerical assessment of the content's sentiment in the range from 0 to 1, where 0 is completely negative and 1 is completely positive.</p><p>Step 2 takes place in parallel with the first step and consists in assessing the frequency (TF) of the occurrence of offensive language, which will be determined by comparing the content of the message with the dictionary of offensive language. The frequency of offensive language will be calculated using the formula <ref type="bibr" target="#b23">[24]</ref>:</p><formula xml:id="formula_0">𝑇𝐹 = 𝑂 𝑤 𝑇𝑜𝑡𝑎𝑙𝐶𝑜𝑢𝑛𝑡 ,<label>(1)</label></formula><p>where 𝑂 𝑤 is the number of offensive words contained in the dictionary, 𝑇𝑜𝑡𝑎𝑙𝐶𝑜𝑢𝑛𝑡 is the total number of words in the analyzed Internet content. The result of this step is a numerical assessment of the concentration of offensive components.</p><p>At step 3, the offensiveness score of the textual content is calculated based on the tone of the sentiment and the concentration of offensive components. The numerical assessment of the offensiveness of the textual content will be calculated using the formula:</p><formula xml:id="formula_1">𝐴𝑏𝑢𝑠𝑖𝑣𝑒𝐿𝑒𝑣𝑒𝑙 = 𝑘 • 𝑇𝐹 + (1 − 𝑘)(1 − 𝑆𝑒𝑛𝑡𝑖𝑚𝑒𝑛𝑡𝑉𝑎𝑙) 2 , (<label>2</label></formula><formula xml:id="formula_2">)</formula><p>where 𝑇𝐹 is a numerical estimate of the concentration of offensive components, 𝑆𝑒𝑛𝑡𝑖𝑚𝑒𝑛𝑡𝑉𝑎𝑙 is a neural network estimate of the sentiment tone of the post, 𝑘 is the coefficient of the threshold of sensitivity of offensive words to the overall level of detection of offensive speech. It is selected according to the policy of the social service under study.</p><p>Accordingly, the output of the method of detecting offensive speech for the Ukrainian language using a recurrent neural network is a numerical assessment of the offensiveness of textual Internet content.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.3.">Formation of a trained RNN sentiment analysis model for detecting offensive speech</head><p>Since one of the inputs to the method of detecting offensive speech for the Ukrainian language using RNN is a trained model, it is necessary to obtain it for Ukrainian-language data. The neural network model is trained according to the algorithm shown in Figure <ref type="figure" target="#fig_2">3</ref>. The next step was to select the neural network architecture. It was decided to use a simple three-layer architecture with an Embedding Layer, LSTM Layer, and Dense Layer with a sigmoidal activation function.</p><p>The next step was to train the neural network with the above architecture. The training stage was conducted in conjunction with the stage of further model evaluation based on such metrics as accuracy, recall, f1, and confusion matrix <ref type="bibr" target="#b24">[25]</ref>.</p><p>Accuracy is defined as the ratio of the number of correctly classified examples to the total number of examples <ref type="bibr" target="#b25">[26]</ref>. Recall is defined as the ratio of the number of correctly classified positive examples to the total number of positive examples. F1-score is calculated as a balanced average between accuracy and precision <ref type="bibr" target="#b26">[27]</ref>.</p><p>The following input parameters were analyzed for the training process: Batch size, number of epochs. The number of training epochs shows how many times the model is to be trained. Batch size shows the number of training examples used within one iteration of neural network training. It is very difficult to immediately determine what the perfect batch size is for the needs of a particular task <ref type="bibr" target="#b27">[28]</ref>, so this parameter will be selected experimentally. The statistics of the metrics for the conducted training are shown in Table <ref type="table" target="#tab_0">1</ref>. Figures <ref type="figure" target="#fig_5">4-7</ref> illustrate the pattern confusions of neural network models. Green is for true positives, red is for true negatives, yellow is for false positives, and blue is for false negatives.    As can be seen from Figures <ref type="figure" target="#fig_5">4-7</ref>, in general, all models cope with the task, but given the purpose of the study, it was decided to use the V5 model, which has the highest Accuracy. Although the V7 model also has fairly high Recall and F1values, it is more important to identify negative samples more accurately.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.4.">Study of the efficiency of the proposed method</head><p>To study the efficiency of the method of detecting offensive speech for the Ukrainian language using a recurrent neural network, a corresponding software implementation was created. Python tools were used for development, and the «wx» library was used for the user interface <ref type="bibr">[29]</ref>. The Sklearn library was used for training and further use of the neural network <ref type="bibr" target="#b28">[30]</ref>. An example of content identification is illustrated in Figure <ref type="figure">8</ref> and Figure <ref type="figure">9</ref>. In the illustrated examples, 2 comments from Facebook were taken as test content. Comment 1: «Original: За півтора року в дворі можна було по бункеру зробити, а не вот це вот все. Що ви приведете до ладу? Подвали? Де в місті є нормальні укриття? / English translate: In a year and a half, you could have built a bunker in the yard, not this crap. What will you improve? Basements? Where are there normal shelters in the city?». This comment was classified by the neural network as negative with a score of 0.303. However, despite the negative content, the comment does not contain direct abuse or offensive words, so the overall offensiveness score is 0.244 with a sensitivity threshold of 0.4. Therefore, this comment is acceptable. Comment 2: «Original: Ви кончені ???? Ми самі облаштовували собі підвал. А вони гроші пиздять далі! / English translate: You're fucking nuts????We were setting up our basement ourselves. And they keep fucking with the money!». As for this comment, it already contains a direct abusive content, is negative with a score of 0.068 (where 0 is completely negative content) and with a content abusiveness score of 0.15, its overall offensive speech score is 0.43, with a threshold value of 0.4, so this review is unacceptable.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Result and Discussion</head><p>Method for detecting offensive speech for the Ukrainian language was developed using a recurrent neural network, which includes a combined approach: an RNN network for determining the numerical assessment of the sentiment tone of the content and an approach for numerically assessing the concentration of offensive components based on the «AbusiveLanguageDataset». The values of metrics for trained versions of the neural networks at 20 epochs and different batch sizes are shown in Figure <ref type="figure" target="#fig_0">10</ref>. The values of metrics for trained versions of neural networks at 10 epochs and different batch sizes are shown in Figure <ref type="figure" target="#fig_0">11</ref>. As can be seen from the diagrams above, the metrics do not decrease below 94%, so the neural network shows high performance in all models for classifying the sentiment tone of text content.</p><p>As part of the study, we collected 50 comments that are not included in the training and test data from Facebook and Instagram, which were evaluated by experts into 2 categories: «acceptable» and «unacceptable». According to the expert evaluation, 32 comments were labeled as «unacceptable» and 18 as «acceptable». Based on the testing conducted using the developed method, 30 comments were unacceptable and 20 were acceptable. The data are illustrated in Table <ref type="table">2</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 2</head><p>Results of testing the method of detecting offensive speech for the Ukrainian language, pcs.</p><p>Acceptable Unacceptable Acceptable 18 0 Unacceptable 2 30</p><p>However, the controversial comments that were characterized by the previous expert as unacceptable, and by the developed method as acceptable, were proposed to be evaluated by 3 more experts, and in the first case, 2 out of 3 also classified them as acceptable. The text of the comment was as follows: «Original: Та я як молодий інженер електрик змушений роботами нижчого рівня перебиватися і по крупинкам практику збирати, бо всілякі там рішали хочуть на своїх підприємствах бачити спеціалістів за максимум 12к гривень. Думаю, я не один такий. / English translate: But as a young electrical engineer, I have to do lower-level jobs and collect my practice bit by bit, because all sorts of bastards want specialists for a maximum of 12 thousand hryvnias at their enterprises. I think I'm not the only one.». The score of this comment by the method of detecting offensive speech for the Ukrainian language was 0.386, with a threshold value of 0.4.</p><p>The second comment was as follows: «Original: Це ж виходить можна безкарно гасити всіх в кого не має родичів. Нема родичів, нема кому звинувачивати. / English translate: That means you can put out everyone who has no relatives with impunity. No relatives, no one to blame.». The content acceptability score was 0.39, and the post was identified as negative with a score of 0.06, where 0 is completely negative content. For this comment, 2 out of three experts gave the rating «unacceptable».</p><p>The results with examples of how the method was used are also shown in Figure <ref type="figure" target="#fig_8">12</ref>. For comparison, the same comments were submitted for the GPT chat evaluation, where he gave the scores shown in Figure <ref type="figure" target="#fig_9">13</ref>.  Comment: «Original: Куропатка яка бігла під автівку, мабуть у неї є запасне життя. / Translation: A dumbass that ran in front of a car, apparently it has a spare life» is rated 0.5 by GPT chat, while the proposed method is rated 0.35. However, the reason for this assessment by GPT chat is the presence of an unfavorable or cruel situation, such as an animal running in front of a car. This can cause negative emotions or seem offensive to those who care about animal welfare. Therefore, this assessment is not entirely correct. The rest of the comments are not in conflict and are determined to be unacceptable by both the proposed method and the GPT chat.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Conclusion</head><p>The paper considers the current state of the field of offensive speech detection, which is one of the key areas in working with texts for many languages. According to the analysis, the main approaches to solving the problem of detecting offensive speech and identifying offensive content were identified, including: sentiment analysis, detection of abusive speech (dictionary approach) and the use of machine learning. It was decided to use a combined approach based on sentiment tone analysis and abusive speech detection (dictionary approach). Also, taking into account the specifics of the application for the Ukrainian language, it was necessary to create an appropriate dataset consisting of tweets from the «Ukrainian Twitter Corpus», marked responses from the «Abusive Language Dataset» and «hotline»</p><p>The tweets and reviews were filtered, and tweets with less than 3 words were removed. After applying restrictions to the tweets and Ukrainian-language reviews from the hotline, their number was as follows: the tweet set was 2400 tweets (1200 positive and 1200 negative), the data set from the hotline was 2000 positive and 2000 negative, the set of posts and comments from the «Facebook» social network was 500 units (250 positive and 250 negative). The number of offensive words in the dictionary was 959 offensive words. In total, the sample amounted to 6900 Ukrainian-language texts.</p><p>To train the RNN, it was decided to divide the dataset in a 60:40 ratio, where 60% is the training sample and 40% is the test sample. Accordingly, the number of training examples was 4140, and the number of test examples was 2760.</p><p>The trained RNN model, which was subsequently used for sentiment analysis, had an accuracy of 0.96, while Recall and F1 had scores of 0.959 and 0.957, respectively.</p><p>Abusive speech detection method for the Ukrainian language using a recurrent neural network was tested on the developed software, and the study shows that the method has a high efficiency of detecting abusive Ukrainian-language content. According to the research, the method has an estimated identification accuracy of more than 90%, however, the assessment of abusiveness can be subjective, and the perception of content can vary from person to person. However, to improve the result, it is necessary to supplement the dictionary of abusive expressions, as the Ukrainian language is rich in surzhik and other foreign language twists,</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Data components of the method for detecting offensive speech for the Ukrainian language</figDesc><graphic coords="4,103.60,433.09,401.30,152.24" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: Scheme of the method for detecting offensive speech for the Ukrainian language using a recurrent neural network</figDesc><graphic coords="5,98.05,293.66,413.10,377.99" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Figure 3 :</head><label>3</label><figDesc>Figure 3: Stages of forming trained RNN sentiment analysis model for detecting offensive speech</figDesc><graphic coords="6,92.73,478.07,423.72,267.65" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Figure 4 :</head><label>4</label><figDesc>Figure 4: Classification of reviews by the V1 model</figDesc><graphic coords="7,72.00,407.30,222.40,131.80" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_4"><head>Figure 5 :Figure 6 :</head><label>56</label><figDesc>Figure 5: Classification of responses by the V2 model</figDesc><graphic coords="7,302.30,407.30,223.78,131.60" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_5"><head>Figure 7 :</head><label>7</label><figDesc>Figure 7: Classification of responses by the V7 model</figDesc><graphic coords="7,302.30,577.98,224.74,148.45" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_6"><head>Figure 8 : 9 :</head><label>89</label><figDesc>Figure 8: Acceptable-negative feedback Figure 9: Unacceptable negative feedback</figDesc><graphic coords="8,74.40,460.34,213.79,265.05" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_7"><head>Figure 10 :Figure 11 :</head><label>1011</label><figDesc>Figure 10: The value of metrics for determining the sentiment of RNN content for 20 epochs</figDesc><graphic coords="9,72.13,397.12,450.72,156.25" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_8"><head>Figure 12 :</head><label>12</label><figDesc>Figure 12: The result of the applied use of the method</figDesc><graphic coords="10,72.00,570.71,450.00,147.00" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_9"><head>Figure 13 :</head><label>13</label><figDesc>Figure 13: Evaluating comments with GPT chat</figDesc><graphic coords="11,72.00,72.00,450.75,173.25" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1 RNN training parameters and results</head><label>1</label><figDesc></figDesc><table><row><cell>Parameters</cell><cell>V1</cell><cell>V2</cell><cell>V3</cell><cell>V4</cell><cell>V5</cell><cell>V6</cell><cell>V7</cell></row><row><cell>Number of</cell><cell>20</cell><cell>20</cell><cell>20</cell><cell>20</cell><cell>10</cell><cell>10</cell><cell>10</cell></row><row><cell>learning epochs</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell>Batch size</cell><cell>128</cell><cell>64</cell><cell>32</cell><cell>16</cell><cell>64</cell><cell>32</cell><cell>16</cell></row><row><cell>Results</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell>Training time (sec)</cell><cell>257</cell><cell>343</cell><cell>503</cell><cell>921</cell><cell>183</cell><cell>255</cell><cell>442</cell></row><row><cell>Accuracy</cell><cell>0.951</cell><cell>0.951</cell><cell>0.957</cell><cell>0.949</cell><cell>0,96</cell><cell>0.956</cell><cell>0.947</cell></row><row><cell>Recall</cell><cell>0.963</cell><cell>0.968</cell><cell>0.948</cell><cell>0.936</cell><cell>0.959</cell><cell>0.943</cell><cell>0.978</cell></row><row><cell>F 1</cell><cell>0.959</cell><cell>0.961</cell><cell>0.956</cell><cell>0.95</cell><cell>0.957</cell><cell>0.956</cell><cell>0.960</cell></row><row><cell>True positive</cell><cell>0.96</cell><cell>0.97</cell><cell>0.95</cell><cell>0.94</cell><cell>0.96</cell><cell>0.94</cell><cell>0.98</cell></row><row><cell>True negative</cell><cell>0.96</cell><cell>0.96</cell><cell>0.97</cell><cell>0.97</cell><cell>0.97</cell><cell>0.98</cell><cell>0.95</cell></row><row><cell>False positive</cell><cell>0.036</cell><cell>0.037</cell><cell>0.028</cell><cell>0.026</cell><cell>0.035</cell><cell>0.02</cell><cell>0.047</cell></row><row><cell>False negative</cell><cell>0.037</cell><cell>0.032</cell><cell>0.05</cell><cell>0.06</cell><cell>0.04</cell><cell>0.06</cell><cell>0.022</cell></row></table></figure>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><p>which are not currently fully represented in the dictionary. The proposed method also has a number of limitations: it works with text content of 3 words or less and no more than 500 words, and it works only in the Ukrainian language.</p><p>Further research can be directed towards practical applications, which can be a useful tool for assessing the level of social acceptability of digital textual content published on social networks and for preventing the spread of harmful or offensive information.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<ptr target="https://transparency.fb.com/policies/community-standards/" />
		<title level="m">Meta, Facebook Community Standards</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Detection and moderation of detrimental content on social media platforms: current status and future directions</title>
		<author>
			<persName><forename type="first">V</forename><forename type="middle">U</forename><surname>Gongane</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">V</forename><surname>Munot</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">D</forename><surname>Anuse</surname></persName>
		</author>
		<idno type="DOI">10.1007/s13278-022-00951-3</idno>
	</analytic>
	<monogr>
		<title level="j">Soc. Netw. Anal. Min</title>
		<imprint>
			<biblScope unit="volume">12</biblScope>
			<biblScope unit="page">129</biblScope>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Social Media and Morality</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">J</forename><surname>Van Bavel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">E</forename><surname>Robertson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Rosario</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Rasmussen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Rathje</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Annual Review of Psychology</title>
		<editor>Van Bavel</editor>
		<imprint>
			<biblScope unit="volume">75</biblScope>
			<biblScope unit="page" from="311" to="340" />
			<date type="published" when="2024">2024. 2024</date>
		</imprint>
	</monogr>
	<note>Annual Review of Psychology</note>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Visual Analytics-Based Method for Sentiment Analysis of COVID-19 Ukrainian Tweets</title>
		<author>
			<persName><forename type="first">O</forename><surname>Kovalchuk</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Slobodzian</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Sobko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Molchanova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Mazurets</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Barmak</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Krak</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Savina</surname></persName>
		</author>
		<idno type="DOI">10.1007/978-3-031-16203-9_33</idno>
		<ptr target="https://link.springer.com/chapter/10.1007/978-3-031-16203-9_33.doi:10.1007/978-3-031-16203-9_33" />
	</analytic>
	<monogr>
		<title level="j">Book Chapter. Lecture Notes on Data Engineering and Communications Technologies</title>
		<imprint>
			<biblScope unit="volume">149</biblScope>
			<biblScope unit="page" from="591" to="607" />
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">E-word of mouth sentiment analysis for user behavior studies</title>
		<author>
			<persName><forename type="first">H</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Zh</forename><surname>Qi Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Zhong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Gong</surname></persName>
		</author>
		<author>
			<persName><surname>Han</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.ipm.2021.102784</idno>
	</analytic>
	<monogr>
		<title level="j">Information Processing &amp; Management</title>
		<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Using Twitter for sentiment analysis towards AstraZeneca/Oxford, Pfizer/BioNTech and Moderna COVID-19 vaccines</title>
		<author>
			<persName><forename type="first">R</forename><surname>Marcec</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Likic</surname></persName>
		</author>
		<idno type="DOI">10.1136/postgradmedj-2021-140685</idno>
	</analytic>
	<monogr>
		<title level="j">Postgraduate Medical Journal</title>
		<imprint>
			<biblScope unit="volume">98</biblScope>
			<biblScope unit="page" from="544" to="550" />
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Covid-19 vaccine hesitancy: Text mining, sentiment analysis and machine learning on COVID-19 vaccination Twitter dataset</title>
		<author>
			<persName><forename type="first">M</forename><surname>Qorib</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Oladunni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Denis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Ososanya</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Cotae</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.eswa.2022.118715</idno>
	</analytic>
	<monogr>
		<title level="j">Expert Systems with Applications</title>
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">An approach to automatic classification of hate speech in sports domain on social media</title>
		<author>
			<persName><forename type="first">V</forename><surname>Stanković</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Mladenović</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename></persName>
		</author>
		<idno type="DOI">10.1186/s40537-023-00766-9</idno>
	</analytic>
	<monogr>
		<title level="j">J Big Data</title>
		<imprint>
			<biblScope unit="volume">10</biblScope>
			<biblScope unit="page">109</biblScope>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Deep Multilayer Neural Network for Predicting the Winner of Football Matches</title>
		<author>
			<persName><forename type="first">S</forename><surname>Anfilets</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Bezobrazov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Golovko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Sachenko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Komar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Dolny</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Kasyanik</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Bykovyy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Mikhno</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Osolinskyi</surname></persName>
		</author>
		<idno type="DOI">10.47839/ijc.19.1.1695</idno>
	</analytic>
	<monogr>
		<title level="j">International Journal of Computing</title>
		<imprint>
			<biblScope unit="volume">19</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="70" to="77" />
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">CIT Kokrajhar Team: LSTM based Deep RNN Architecture for Hate Speech and Offensive Content</title>
		<author>
			<persName><forename type="first">B</forename><forename type="middle">N</forename><surname>Saha</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Senapati</surname></persName>
		</author>
		<ptr target="https://ceur-ws.org/Vol-2517/T3-24.pdf" />
	</analytic>
	<monogr>
		<title level="m">HASOC) Identification in Indo-European Languages</title>
				<meeting><address><addrLine>FIRE; Kolkata, India</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2019-12">2019. December. 2019</date>
			<biblScope unit="page" from="12" to="15" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">An RNN-based Hybrid Model for Classification of Electrooculogram Signal for HCI</title>
		<author>
			<persName><forename type="first">K</forename><forename type="middle">S</forename><surname>Roy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">M R</forename><surname>Islam</surname></persName>
		</author>
		<idno type="DOI">10.47839/ijc.22.3.3228</idno>
	</analytic>
	<monogr>
		<title level="j">International Journal of Computing</title>
		<imprint>
			<biblScope unit="volume">22</biblScope>
			<biblScope unit="issue">3</biblScope>
			<biblScope unit="page" from="335" to="344" />
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<monogr>
		<ptr target="https://time.com/5739688/facebook-hate-speech-languages/" />
		<title level="m">Time, Facebook Says It&apos;s Removing More Hate Speech Than Ever Before</title>
				<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
	<note>But There&apos;s a Catch</note>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">What Teams Do: Exploring Volunteer Content Moderation Team Labor on Facebook</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">D</forename><surname>Gibson</surname></persName>
		</author>
		<idno type="DOI">10.1177/20563051231186109</idno>
	</analytic>
	<monogr>
		<title level="j">Social Media + Society</title>
		<imprint>
			<biblScope unit="volume">9</biblScope>
			<biblScope unit="issue">3</biblScope>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Social media mindsets: a new approach to understanding social media use and psychological well-being</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">Y</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">T</forename><surname>Hancock</surname></persName>
		</author>
		<idno type="DOI">10.1093/jcmc/zmad048</idno>
		<ptr target="https://academic.oup.com/jcmc/article/29/1/zmad048/7612379.doi:10.1093/jcmc/zmad048" />
	</analytic>
	<monogr>
		<title level="j">Journal of Computer-Mediated Communication</title>
		<imprint>
			<biblScope unit="volume">29</biblScope>
			<biblScope unit="issue">1</biblScope>
			<date type="published" when="2024-01">January, 2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Hate crimes in social media: A criminological review</title>
		<author>
			<persName><forename type="first">N</forename><surname>Yesmen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Arnab</surname></persName>
		</author>
		<author>
			<persName><surname>Sumona Sh</surname></persName>
		</author>
		<idno type="DOI">10.53625/ijss.v3i1.5607</idno>
	</analytic>
	<monogr>
		<title level="j">International Journal of Social Science</title>
		<imprint>
			<biblScope unit="volume">3</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="23" to="28" />
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Morality in social media: A scoping review</title>
		<author>
			<persName><forename type="first">D</forename><surname>Neumann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Rhodes</surname></persName>
		</author>
		<idno type="DOI">10.1177/14614448231166056</idno>
		<ptr target="https://journals.sagepub.com/doi/full/10.1177/14614448231166056.doi:10.1177/14614448231166056" />
	</analytic>
	<monogr>
		<title level="j">New Media &amp; Society</title>
		<imprint>
			<biblScope unit="volume">26</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page" from="1096" to="1126" />
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Online Content Moderation: Does Justice Need a Human Face?</title>
		<author>
			<persName><forename type="first">M</forename><surname>Katsaros</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Kim</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Tyler</surname></persName>
		</author>
		<idno type="DOI">10.1080/10447318.2023.2210879</idno>
		<ptr target="https://www.tandfonline.com/doi/abs/10.1080/10447318.2023.2210879.doi:10.1080/10447318.2023.2210879" />
	</analytic>
	<monogr>
		<title level="j">International Journal of Human-Computer Interaction</title>
		<imprint>
			<biblScope unit="volume">40</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="66" to="77" />
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Speech Emotion Recognition using Hybrid Architectures</title>
		<author>
			<persName><forename type="first">M</forename><surname>Norval</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Wang</surname></persName>
		</author>
		<idno type="DOI">10.47839/ijc.23.1.3430</idno>
	</analytic>
	<monogr>
		<title level="j">International Journal of Computing</title>
		<imprint>
			<biblScope unit="volume">23</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="1" to="10" />
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<monogr>
		<ptr target="https://github.com/saganoren/ukr-twi-corpus" />
		<title level="m">Ukr-twi-corpus</title>
				<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<monogr>
		<author>
			<persName><surname>Hatespeechdata</surname></persName>
		</author>
		<ptr target="https://hatespeechdata.com/" />
		<title level="m">Hate Speech Dataset Catalogue</title>
				<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">Method for Sentiment Analysis of Ukrainian-Language Reviews in E-Commerce Using RoBERTa Neural Network</title>
		<author>
			<persName><forename type="first">O</forename><surname>Zalutska</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Molchanova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Sobko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Mazurets</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Pasichnyk</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Barmak</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Krak</surname></persName>
		</author>
		<ptr target="https://ceur-ws.org/Vol-3387/paper26.pdf" />
	</analytic>
	<monogr>
		<title level="m">CEUR Workshop Proceedings</title>
				<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="volume">3387</biblScope>
			<biblScope unit="page" from="561" to="571" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">More than a Feeling: Accuracy and Application of Sentiment Analysis</title>
		<author>
			<persName><forename type="first">J</forename><surname>Hartmann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Heitmann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ch</forename><surname>Siebert</surname></persName>
		</author>
		<author>
			<persName><surname>Ch</surname></persName>
		</author>
		<author>
			<persName><surname>Schamp</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.ijresmar.2022.05.005</idno>
	</analytic>
	<monogr>
		<title level="j">International Journal of Research in Marketing</title>
		<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">How do text characteristics impact user engagement in social media posts: Modeling content readability, length, and hashtags number in Facebook</title>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">C</forename><surname>Gkikas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Tzafilkou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">K</forename><surname>Theodoridis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Garmpis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Gkikas</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.jjimei.2022.100067</idno>
	</analytic>
	<monogr>
		<title level="j">International Journal of Information Management Data Insights.тVolume</title>
		<imprint>
			<biblScope unit="volume">2</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">A machine learning tool for collecting and analyzing subjective road safety data from Twitter</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">M</forename><surname>Abedi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Sacchi</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.eswa.2023.122582</idno>
		<ptr target="https://www.sciencedirect.com/science/article/abs/pii/S0957417423030841.doi:10.1016/j.eswa.2023.122582" />
	</analytic>
	<monogr>
		<title level="j">Expert Systems with Applications</title>
		<imprint>
			<biblScope unit="volume">240</biblScope>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<analytic>
		<title level="a" type="main">The practice implementation of the information technology for automated definition of semantic terms sets in the content of educational materials</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Krak</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Barmak</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Mazurets</surname></persName>
		</author>
		<idno type="DOI">10.15407/pp2018.02.245</idno>
	</analytic>
	<monogr>
		<title level="m">CEUR Workshop Proceedings</title>
				<imprint>
			<date type="published" when="2018">2018</date>
			<biblScope unit="volume">2139</biblScope>
			<biblScope unit="page" from="245" to="254" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b25">
	<analytic>
		<title level="a" type="main">More than a Feeling: Accuracy and Application of Sentiment Analysis</title>
		<author>
			<persName><forename type="first">J</forename><surname>Hartmann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Heitmann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ch</forename><surname>Siebert</surname></persName>
		</author>
		<author>
			<persName><surname>Ch</surname></persName>
		</author>
		<author>
			<persName><surname>Schamp</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.ijresmar.2022.05.005</idno>
	</analytic>
	<monogr>
		<title level="j">International Journal of Research in Marketing</title>
		<imprint>
			<biblScope unit="volume">40</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="75" to="87" />
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b26">
	<monogr>
		<author>
			<persName><surname>Scikit-Learn</surname></persName>
		</author>
		<idno>f1_score. 2023</idno>
		<ptr target="https://scikit-learn.org/stable/modules/generated/sklearn.metrics.f1_score.html" />
		<title level="m">sklearn.metrics</title>
				<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b27">
	<monogr>
		<title level="m" type="main">How does Batch Size impact your model learning</title>
		<author>
			<persName><surname>Medium</surname></persName>
		</author>
		<ptr target="https://medium.com/geekculture/how-does-batch-size-impact-your-model-learning-2dd34d9fb1fa" />
		<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b28">
	<monogr>
		<title level="m" type="main">Machine Learning in Python</title>
		<ptr target="https://scikit-learn.org/stable/" />
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
	<note>Scikit-learn</note>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
