<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>A. Toktarova);</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Hate speech detection on social media using machine learning</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Baktykul Jakhanova</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Aigerim Toktarova</string-name>
          <xref ref-type="aff" rid="aff3">3</xref>
          <xref ref-type="aff" rid="aff4">4</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Aigerim Altayeva</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Rustam Abdrakhmanov</string-name>
          <email>rustam.abdyrakhmanov@gmail.com</email>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Danyar Sultan</string-name>
          <email>daniyarsultan916@gmail.com</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Asfendiyarov Kazakh National Medical University</institution>
          ,
          <addr-line>050000, Almaty</addr-line>
          ,
          <country country="KZ">Kazakhstan</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>International Information Technology University</institution>
          ,
          <addr-line>34/1 Manas St., Almaty, 050000</addr-line>
          ,
          <country country="KZ">Kazakhstan</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>International University of Tourism and Hospitality</institution>
          ,
          <addr-line>Turkistan, 161200</addr-line>
          ,
          <country country="KZ">Kazakhstan</country>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>Khoja Akhmet Yassawi International Kazakh - Turkish University</institution>
          ,
          <addr-line>Turkistan, 161200</addr-line>
          ,
          <country country="KZ">Kazakhstan</country>
        </aff>
        <aff id="aff4">
          <label>4</label>
          <institution>M. Auezov South Kazakhstan University</institution>
          ,
          <addr-line>Shymkent, 160000</addr-line>
          ,
          <country country="KZ">Kazakhstan</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>1923</year>
      </pub-date>
      <volume>000</volume>
      <fpage>0</fpage>
      <lpage>0002</lpage>
      <abstract>
        <p>This research study uses a thorough analysis of numerous machine learning and deep learning techniques to address the crucial problem of cyberbullying detection in the social media arena. By carefully assessing these methods using industry-standard measures including F-measure, AUC-ROC, precision, accuracy, and recall, the study investigates how effective these approaches are. The outcomes show how well deep learning models-specifically, the bidirectional long-short-term memory (BiLSTM) architecture-perform, consistently surpassing other techniques on a range of categorization tasks. Confusion matrices and graphical depictions provide more insight into the model's functionality, showcasing the extraordinary capacity of the BiLSTM-based model to correctly identify and categorize instances of cyberbullying. The significance of sophisticated neural network architectures in identifying the intricacy of hateful and objectionable content on the internet underscores by these findings. This study offers insightful information for encouraging early detection and mitigation of cyberbullying, which in turn promotes secure and welcoming online communities. Future studies could look into real-time detection systems, hybrid techniques, or the integration of complementing elements to further develop and enhance cutting-edge technology in tackling this significant social issue.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Machine learning</kwd>
        <kwd>deep learning</kwd>
        <kwd>hate speech</kwd>
        <kwd>CNN</kwd>
        <kwd>RNN</kwd>
        <kwd>LSTM</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>Hate speech has spread as a result of this accessibility and openness, though. The identification
and mitigation of hate speech has gained popularity, leading scholars to investigate several
methodologies like deep learning (DL) and machine learning (ML).</p>
      <p>
        Hate can lead to violence, social discord, and psychological injury to the targeted individuals or
communities. It typified by offensive, damaging, or discriminating content [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. Its existence on social
media not only jeopardizes user safety but also damages the platforms' credibility and reputation.
      </p>
      <p>
        Furthermore, these techniques demonstrate exceptional proficiency in managing the rapid and
intricate nature of social media writing, which distinguished by its brevity, casual tone, and frequent
spelling errors. They have the ability to efficiently handle and examine textual information from
diverse origins, such as tweets, comments, and forum postings [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ].
      </p>
      <p>Recent research has demonstrated encouraging outcomes in the identification of hate speech
through the utilization of machine learning (ML) and deep learning (DL) methodologies. These
methods have successfully attained high levels of accuracy, precision, and recall rates, presenting a
promising answer to the persistent problem of controlling hate speech on the internet [7-9].</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related works</title>
      <p>Cyberbullying can transpire continuously, providing no opportunity for individuals to feel secure;
messages and comments may arrive unexpectedly at any moment, resulting in significant
psychological effects on adolescents. Furthermore, the anonymity of the Internet may prevent the
teenager from identifying the one perpetrating the bullying, potentially exacerbating their dread. In
contrast to physical violence, the repercussions of emotional abuse ultimately impact psychological
well-being. Identifying a victim of emotional abuse is challenging. The automatic detection of
cyberbullying can prevent it promptly [7-12].</p>
      <p>
        As a result, social media corporations and policymakers have taken proactive steps to combat the
dissemination of hate speech [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ].
      </p>
      <p>
        Machine learning techniques utilize natural language processing (NLP) technologies to
automatically detect and categorize hate speech material [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. Significantly, they do not depend
exclusively on conventional methods that use keywords, which frequently fail to identify nuanced
manifestations of hate speech [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ].
      </p>
      <p>
        An internet benefit of utilizing machine learning (ML) and deep learning (DL) techniques in hate
speech identification is their capacity to adapt. Hate speech undergoes a process of development over
time, assimilating novel derogatory terms, symbols, and phrases that may not be effectively identified
by rigid rule-based systems. Machine learning (ML) and deep learning (DL) models have the ability to
learn and adjust to new patterns in a continuous manner [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ].
      </p>
      <p>This work provides a comprehensive overview of several approaches, methodologies, and
datasets employed in hate speech research, emphasizing their respective advantages and
disadvantages [10]. Through a thorough examination of the complexities involved in identifying hate
speech, our objective is to make a valuable contribution to the ongoing discussion on how to tackle
this crucial problem. Additionally, we seek to offer valuable insights that can guide future research
and advancements in this particular domain [11]. In addition, we explore the approaches and
empirical findings, providing a comprehensive examination of the efficacy of machine learning (ML)
and deep learning (DL) methods in identifying hate speech on social media platforms.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Problem statement</title>
      <p>The issue of early detection of cyberbullying within the realm of social networking platforms may
inherently differ from the challenge associated with classifying distinct manifestations of
cyberbullying [12]. In the context delineated herein, we identify a cohort of social media interactions
collectively denoted as "S." Consequently, it becomes plausible that a subset of these interactions may
indeed represent instances of cyberbullying. The progression of such interactions on a given social
network can be succinctly characterized using the following equation (1):</p>
      <p>S ={s1 , s2 , . . . , s|S|}</p>
      <p>Within the scope of this investigation, the variable "S" denotes the aggregate count of sessions,
while the variable "i" signifies the present session under consideration. It is noteworthy that the order
in which submissions occur during a given session can undergo modifications at distinct temporal
junctures, influenced by an array of multifaceted determinants.</p>
      <p>Ps=(⟨ P1S , t1S ⟩ , ⟨ P2S , t2S ⟩ , . . . , ⟨ PnS , t nS ⟩)</p>
      <p>In the context of this study, the tuple denoted as "P" symbolizes the kth post within the context of
the social network session, while "s" corresponds to the timestamp indicating the precise moment at
which post P disseminated.</p>
      <p>Simultaneously, a distinctive vector of attributes harnessed for the unequivocal identification of
each individual post.</p>
      <p>Pk =[ f kS , f kS , . . . , f kS ] , k ∈[ 1 , n ]</p>
      <p>S</p>
      <p>1 2 n</p>
      <p>Hence, the primary aim of this endeavor is to amass the requisite insights, enabling the formulation
of a function denoted as "f," which possesses the capability to discern the association between a given
text and the presence of hate speech.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Materials and methods</title>
      <p>The prototype database for the aforementioned system was established by an examination of 215
English-language Twitter accounts, encompassing a total of 200,000 tweets, of which over 4,000 tweets
were subjected to detailed analysis. Analysis revealed 583 English-language tweets exhibiting
characteristics of the detrimental tactic known as “cyberbullying”. Electronic verbal bullying was
predominantly observed in posts by adolescents aged 11-17 and young adults aged 18-35. Teenage
cyberbullying typically involved groups, while electronic bullying among adolescents followed a "one
bully – one victim" model.</p>
      <p>Illustration of the developed model designed for the classification of hate speech instances is
visually depicted in Figure 1. The model comprises distinct stages, which include preprocessing,
feature extraction, classification, and evaluation. This section entails a comprehensive exploration of
each of these stages, with a deliberate emphasis on the intricacies involved.</p>
      <p>Word2Vec is a widely used feature representation technique in NLP [13]. It belongs to the family
of word embedding methods that transform words into continuous vector representations in a
highdimensional space. Word2Vec captures semantic and contextual relationships between words by
learning from large text corpora [14].</p>
      <p>This technique assigns each word a vector in such a way that words with similar meanings are
closer to each other in the vector space [15]. Word2Vec enhances NLP tasks by enabling models to
understand the context and semantics of words, which is particularly valuable for applications like
sentiment analysis, document clustering, and information retrieval [16]. By converting words into
vectors, Word2Vec contributes to more effective and accurate text analysis and natural language
understanding.</p>
      <p>Bag of Words (BoW) The Bag of Words (BoW) model stands as a foundational technique in the
field of natural language processing (NLP) and text mining, facilitating the transformation of textual
information into numerical data, thereby enabling computational algorithms to process language. This
model operates by constructing a vocabulary of unique words from a corpus and then converting text
documents into vectors, where each vector element represents the frequency of a particular word in
the document [17]. Despite its simplicity, the BoW model has been instrumental in numerous NLP
applications, including document classification, sentiment analysis, and topic modeling [18].
However, it is not without limitations; notably, the model's disregard for word order and context can
lead to a loss of semantic meaning [19]. Furthermore, the high dimensionality of the resulting vectors,
especially with large vocabularies, poses challenges for computational efficiency [20]. Nonetheless,
the BoW model's ease of implementation and interpretability continues to make it a valuable tool in
the initial stages of text analysis projects.</p>
      <sec id="sec-4-1">
        <title>4.1. Machine learning for hate speech detection</title>
        <p>In the realm of hate speech detection within social networks, various machine learning models have
been employed to address the complex task of distinguishing between offensive language and benign
content. Each of these models offers distinct advantages and trade-offs, making them suitable for
different aspects of the problem [21].</p>
        <p>Decision Trees: Decision tree models provide a structured representation of decision-making
processes. They are interpretable and can be valuable for identifying explicit patterns and features
indicative of hate speech [22]. However, they may struggle to capture more subtle contextual cues.</p>
        <p>Logistic Regression allows for the estimation of probabilities and predictions in situations where
the outcome is categorical, such as spam email detection or medical diagnosis. Logistic Regression's
simplicity and interpretability make it a valuable tool in various fields, including data analysis,
healthcare, and marketing.</p>
        <p>Naive Bayes: Naive Bayes models are based on probabilistic principles. They are especially adept at
handling text data due to their independence assumptions. Naive Bayes models can efficiently process
large volumes of text and can adapt well to the high perplexity of social media content.</p>
        <p>K-Nearest Neighbors [24] can be useful for identifying similar posts with similar hate speech
content, yet it may struggle with high-dimensional data.</p>
        <p>Support Vector Machines (SVM) is robust against overfitting and can handle high-dimensional
feature spaces [25]. SVMs can be effective in capturing complex decision boundaries in hate speech
detection.</p>
        <p>The choice of machine learning model should consider the specific characteristics of the hate
speech detection problem, such as the prevalence of subtle hate speech, the dimensionality of the text
data, and the need for interpretability. Often, a combination of these models in ensemble techniques or
hybrid approaches is employed to harness their individual strengths and mitigate their limitations,
ultimately improving the overall performance of hate speech detection systems.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Deep learning for hate speech detection</title>
        <p>In the domain of hate speech detection in social networks, deep learning models have emerged as
potent tools due to their capacity to capture intricate linguistic nuances and contextual dependencies
within textual data. Three prominent deep learning architectures, Convolutional Neural Networks
(CNNs), Long Short-Term Memory networks (LSTMs), and Bidirectional LSTMs (BiLSTMs), have
been widely employed to address the complexities inherent in this task [26].</p>
        <p>Convolutional Neural Networks (CNNs): CNNs, initially designed for image processing, have
been adapted for text analysis (Figure 2). They employ convolutional layers to detect local patterns
and hierarchies of features within text. In hate speech detection, CNNs can effectively identify
significant textual structures and are particularly adept at capturing short-range dependencies such
as n-grams and patterns indicative of hate speech expressions.</p>
        <p>Long Short-Term Memory networks (LSTMs): LSTMs are recurrent neural networks (RNNs)
designed to capture sequential information over longer distances (Figure 3). They excel in modeling
dependencies over time and have proven valuable in understanding the temporal aspects of hate
speech evolution. LSTMs can detect contextually relevant information and provide a dynamic
understanding of text.</p>
        <p>Bidirectional LSTMs (BiLSTMs): BiLSTMs extend the LSTM architecture by processing sequences
in both forward and backward directions, allowing them to capture bidirectional dependencies (Figure
4). In hate speech detection, BiLSTMs are particularly effective in understanding contextual nuances
and capturing relationships between words in both preceding and succeeding contexts.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Experiment results</title>
      <sec id="sec-5-1">
        <title>5.1. Evaluation parameters</title>
        <p>In the context of hate speech detection within social networks, evaluating the performance of
machine learning and deep learning models is crucial for assessing their effectiveness in mitigating
the spread of offensive content. Several evaluation parameters commonly employed to gauge the
performance of such models comprehensively.</p>
        <p>TP +TN
accuracy =</p>
        <p>P + N
preision=
recall=</p>
        <p>TP
TP + FP</p>
        <p>TP</p>
        <p>TP + FN
2× precision×recall
F 1=
precision+recall
(6)
(7)
(8)
(9)</p>
        <p>In the context of hate speech detection, a balance between precision and recall is often sought, as
falsely classifying non-hate speech as hate speech (false positives) or failing to detect hate speech (false
negatives) can have significant real-world consequences. Researchers and practitioners may also
consider domain-specific evaluation metrics and adjust the thresholds based on the desired trade-offs
between precision and recall. Robust evaluation methodologies are essential to developing and
deploying effective hate speech detection systems that contribute to fostering safer and more inclusive
online communities.</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>6. Results</title>
      <p>Evaluation metrics are essential for quantifying the effectiveness of algorithms in classifying instances
within the cyberbullying classification dataset.</p>
      <p>Confusion matrices, as depicted in Figure 5, play a pivotal role in visualizing the outcomes of these
classification techniques. They provide a clear representation of the actual distribution of
classification results across different classes.</p>
      <p>By utilizing confusion matrices, researchers can discern the true positive, true negative, false
positive, and false negative predictions, enabling a comprehensive understanding of the model's
performance in distinguishing between cyberbullying and non-cyberbullying instances. These
evaluations are essential for refining and optimizing cyberbullying detection algorithms to enhance
their accuracy and reliability in addressing the critical issue of online harassment and bullying.</p>
      <p>Figure 6 presents a comparative analysis between the proposed model and a range of other
machine learning and deep learning models employed in this study. The performance evaluation in
each classification scenario is conducted by computing the Area Under the Receiver Operating
Characteristic Curve (AUC-ROC), encompassing all extracted features. This approach allows for a
comprehensive assessment of the discriminatory power and effectiveness of the suggested model in
comparison to alternative methodologies, thereby providing valuable insights into its performance
across different classification tasks.</p>
      <p>Figure 6: Results in Hate Speech Detection</p>
      <p>These findings underscore the efficacy and robustness of the BiLSTM-based model in effectively
discriminating and classifying the target classes, further substantiating the merit of deep learning
paradigms in the context of the study.</p>
    </sec>
    <sec id="sec-7">
      <title>7. Conclusion</title>
      <p>In conclusion, this research paper has delved into the critical realm of cyberbullying detection within
the context of social networks. Through a comprehensive exploration of various machine learning and
deep learning methodologies, coupled with meticulous evaluation using metrics such as Accuracy,
Precision, Recall, F-measure, and AUC-ROC, we have endeavored to shed light on the effectiveness of
these techniques in addressing the m ultifaceted challenge of identifying instances of cyberbullying.</p>
      <p>Our findings underscore the pivotal role that deep learning models, particularly the Bidirectional
Long Short-Term Memory (BiLSTM) architecture, play in enhancing the discriminatory power and
accuracy of cyberbullying detection systems. The consistent superiority of the BiLSTM-based model
across various classification tasks reaffirms the potential of advanced neural network structures in
capturing the intricacies of online hate speech and offensive content. Moreover, the utilization of
confusion matrices and visualizations has allowed for a nuanced understanding of model
performance. This research contributes valuable insights into the ongoing efforts to create safer and
more inclusive online spaces, where the early identification and mitigation of cyberbullying are
paramount. Future research endeavors may explore hybrid approaches, leverage additional features,
or delve into real-time cyberbullying detection systems to further refine and enhance the
state-ofthe-art in this vital domain</p>
    </sec>
    <sec id="sec-8">
      <title>Acknowledgements</title>
      <p>This work was supported by the research project ― Automatic detection of cyberbullying among
young people in social networks using artificial intelligence funded by the Ministry of Science and
Higher Education of the Republic of Kazakhstan. Grant No. IRN AP23488900.</p>
    </sec>
    <sec id="sec-9">
      <title>Declaration on Generative AI</title>
      <p>The authors have not employed any Generative AI tools.
[7] A. Al-Marghilani, "Artificial intelligence-enabled cyberbullying-free online social networks in
smart cities," International Journal of Computational Intelligence Systems, vol. 15, no. 1, pp. 1–
13, 2022.
[8] C. Theng, N. Othman, R. Abdullah, S. Anawar, Z. Ayop et al., “Cyberbullying detection in twitter
using sentiment analysis,” International Journal of Computer Science &amp; Network Security, vol.
21, no. 11, pp. 1-10, 2021.
[9] S. Sadiq, A. Mehmood, S. Ullah, M. Ahmad, G. Choi et al., “Aggression detection through deep
neural model on twitter,” Future Generation Computer Systems, vol. 114, no. 1, pp. 120–129,
2021.
[10] E. Sarac Essiz and M. Oturakci, "Artificial bee colony–based feature selection algorithm for
cyberbullying," The Computer Journal, vol. 64, no. 3, pp. 305–313, 2021.
[11] C. E. Gomez, M. O. Sztainberg and R. E. Trana, "Curating cyberbullying datasets: a human-AI
collaborative approach," International journal of bullying prevention, vol. 4, no. 1, pp. 35-46,
2022.
[12] S. Salawu, J. Lumsden and Y. He, "A mobile-based system for preventing online abuse and
cyberbullying," International Journal of Bullying Prevention, vol. 4, no. 1, pp. 66–88, 2022.
[13] M. Mladenović, V. Ošmjanski and S. V. Stanković, “Cyber-aggression, cyberbullying, and
cybergrooming: a survey and research challenges,” ACM Computing Surveys (CSUR), vol. 54, no.1, pp.
1–42, 2021.
[14] S. R. Sangwan and M. P. S. Bhatia, “Denigrate comment detection in low-resource Hindi
language using attention-based residual networks,” Transactions on Asian and Low-Resource
Language Information Processing, vol. 21, no. 1, pp. 1–14, 2021.
[15] T. T. Aurpa, R. Sadik and M. S. Ahmed, “Abusive Bangla comments detection on Facebook using
transformer-based deep learning models,” Social Network Analysis and Mining, vol. 12, no.1, pp.
1–14, 2022.
[16] R. Yan, Y. Li, D. Li, Y. Wang, Y. Zhu et al., “A Stochastic Algorithm Based on Reverse Sampling
Technique to Fight Against the Cyberbullying,” ACM Transactions on Knowledge Discovery
from Data (TKDD), vol. 15, no. 4, pp. 1–22, 2021.
[17] C. J. Yin, Z. Ayop, S. Anawar, N. F. Othman and N. M. Zainudin, “Slangs and Short forms of
Malay Twitter Sentiment Analysis using Supervised Machine Learning,” International Journal of
Computer Science &amp; Network Security, vol. 21, no. 11, pp. 294–300, 2021.
[18] G. Jacobs, C. Van Hee and V. Hoste, “Automatic classification of participant roles in
cyberbullying: Can we detect victims, bullies, and bystanders in social media text?,” Natural
Language Engineering, vol. 28, no. 2, pp. 141–166, 2022.
[19] A. Jevremovic, M. Veinovic, M. Cabarkapa, M. Krstic, I. Chorbev et al., “Keeping Children Safe
Online With Limited Resources: Analyzing What is Seen and Heard,” IEEE Access, vol. 9, no. 1,
pp. 132723–132732, 2021.
[20] K. Kumari, J. P. Singh, Y. K. Dwivedi and N. P. Rana, “Multi-modal aggression identification
using convolutional neural network and binary particle swarm optimization,” Future
Generation Computer Systems, vol. 118, no. 1, pp. 187–197, 2021.
[21] A. M. Abbas, “Social network analysis using deep learning: applications and schemes,” Social</p>
      <p>Network Analysis and Mining, vol.11, no. 1, pp. 1–21, 2021.
[22] Toktarova, A., et al. "Automatic offensive language detection in online user generated
contents." Journal of Theoretical and Applied Information Technology 99.9 (2021): 2054-2067.
[23] S. Mohammed, W. C. Fang, A. E. Hassanien and T. H. Kim, “Advanced Data Mining Tools and</p>
      <p>Methods for Social Computing,” The Computer Journal, vol. 64, no. 3, pp. 281–285, 2021.
[24] Baimakhanova, Aigerim, et al. "Automatic Classification of Scanned Electronic University
Documents using Deep Neural Networks with Conv2D Layers." International Journal of
Advanced Computer Science and Applications 14.5 (2023).
[25] Toktarova, A., et al. "Automated Hate Speech Classification using Emotion Analysis in Social</p>
      <p>Media User Generated Texts." J. Theor. Appl. Inf. Technol 100 (2022): 6621-6634.
[26] Makhanova, Zlikha, et al. "A Deep Residual Network Designed for Detecting Cracks in Buildings
of Historical Significance." International Journal of Advanced Computer Science &amp;
Applications 15.5 (2024).</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>T.</given-names>
            <surname>Alsubait</surname>
          </string-name>
          and
          <string-name>
            <given-names>D.</given-names>
            <surname>Alfageh</surname>
          </string-name>
          , “
          <article-title>Comparison of machine learning techniques for cyberbullying detection on youtube arabic comments</article-title>
          ,”
          <source>International Journal of Computer Science and Network Security</source>
          , vol.
          <volume>21</volume>
          , no.
          <issue>1</issue>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>5</lpage>
          ,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>A.</given-names>
            <surname>Dewani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Memon</surname>
          </string-name>
          and
          <string-name>
            <given-names>S</given-names>
            <surname>Bhatti</surname>
          </string-name>
          , “
          <article-title>Cyberbullying detection: Advanced preprocessing techniques &amp; deep learning architecture for roman urdu data</article-title>
          ,
          <source>” Journal of Big Data</source>
          , vol.
          <volume>8</volume>
          , no.
          <issue>1</issue>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>20</lpage>
          ,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>D.</given-names>
            <surname>Hall</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Silva</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wheeler</surname>
          </string-name>
          , L. Cheng and K. Baumel, “
          <article-title>Harnessing the power of interdisciplinary research with psychology-informed cyberbullying detection models</article-title>
          ,”
          <source>International Journal of Bullying Prevention</source>
          , vol.
          <volume>4</volume>
          , no.
          <issue>1</issue>
          , pp.
          <fpage>47</fpage>
          -
          <lpage>54</lpage>
          ,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>K.</given-names>
            <surname>Arce-Ruelas</surname>
          </string-name>
          , “
          <article-title>Automatic cyberbullying detection: A Mexican case in high school and Higher Education Students,” IEEE Latin America Transactions</article-title>
          , vol.
          <volume>20</volume>
          , no.
          <issue>5</issue>
          , pp.
          <fpage>770</fpage>
          -
          <lpage>779</lpage>
          ,
          <year>2022</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>T.</given-names>
            <surname>Ahmed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Rahman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Nur</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Islam</surname>
          </string-name>
          and
          <string-name>
            <surname>D. Das</surname>
          </string-name>
          , “
          <article-title>Natural language processing and machine learning based cyberbullying detection for Bangla and romanized bangla texts,” TELKOMNIKA (Telecommunication Computing Electronics and Control)</article-title>
          , vol.
          <volume>20</volume>
          , no. 1 pp.
          <fpage>89</fpage>
          -
          <lpage>97</lpage>
          ,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <surname>Toktarova</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sultan</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Azhibekova</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          (
          <year>2024</year>
          , May).
          <source>Review of Machine Learning Models in Cyberbullying Detection Problem. In 2024 IEEE 4th International Conference on Smart Information Systems and Technologies (SIST) </source>
          (pp.
          <fpage>233</fpage>
          -
          <lpage>238</lpage>
          ). IEEE.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>