<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <issn pub-type="ppub">1613-0073</issn>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Analysis for Hate Speech and Ofensive Language in Low-Resource Indo-Aryan Languages</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Ch Muhammad Awais</string-name>
          <email>c.awais@studenti.unipi.it</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Jayveersinh Raj</string-name>
          <email>j.raj@innopolis.university</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="editor">
          <string-name>Multilingual Toxicity Analysis, Hate Speech Detection, Low-Resource Languages, Natural Language</string-name>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Innopolis University</institution>
          ,
          <addr-line>Innopolis</addr-line>
          ,
          <country country="RU">Russia</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>University of Pisa</institution>
          ,
          <addr-line>Pisa</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>less than 6th. respectively). In our interconnected digital landscape, tackling the proliferation of hate speech and ofensive content across languages is a pressing challenge. This paper presents a pioneering approach, fine-tuning the XLM-RoBERTa model for hate speech detection in low-resource languages, transcending linguistic boundaries. Our work is motivated by our participation in the Hate Speech and Ofensive Content Identification (HASOC) competition, where our team, AI Alchemists, achieved 3rd place in Sinhala, and participated in Task 1 and 4 with the provided datasets. Our average position was 4th, with no position Our experiments show that the model can be used to detect hate speech in many diferent languages, including Sinhala and Bodo, with high accuracy (F1 scores of 0.8345 and 0.844, respectively). It also achieved promising results on Gujarati, Bengali, and Assamese (F1 scores of 0.793, 0.726, and 0.707, We emphasize that the quality and size of the training dataset are important factors in the performance of the model. With further research and access to more balanced datasets, we believe that our model has the potential to outperform state-of-the-art models and curb hate speech across diverse linguistic landscapes, fostering more inclusive online spaces.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>CEUR
ceur-ws.org</p>
    </sec>
    <sec id="sec-2">
      <title>1. Introduction</title>
      <p>
        In today’s digital age, the widespread proliferation of ofensive content has become an urgent
societal concern. Hate speech, ofensive language, and profanity have found fertile ground
on various online platforms, perpetuating hostility and division among users [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. Efectively
addressing this issue requires advanced tools and methodologies capable of identifying and
mitigating ofensive content with precision since it is challenging to categorize tweets or comments
without obvious hateful keyword [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], and partly because there isn’t a consensus definition of
hate speech, examination of its demographic influences, or investigation of its most potent
https://cm-awais.github.io/ (C. M. Awais); https://jayveersinh-raj.github.io/ (J. Raj)
      </p>
      <p>
        © 2023 Copyright for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0).
CEUR
Workshop
Proceedings
characteristics [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. Natural Language Processing (NLP) has emerged as a powerful approach in
this endeavor, harnessing the fusion of machine learning and language analysis to discern and
counteract toxic language [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ].
      </p>
      <p>
        NLP techniques have proven instrumental in analyzing vast volumes of textual data, enabling
the detection of ofensive content across diverse linguistic landscapes. By leveraging
computational linguistics and statistical models, NLP algorithms can automatically classify texts into
ofensive and non-ofensive categories [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. The ability to aggregate feedback without human
intervention makes this incredibly helpful [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. These algorithms excel at recognizing patterns,
contextual cues, and linguistic features associated with ofensive language, thereby aiding in
the identification and moderation of problematic content [
        <xref ref-type="bibr" rid="ref7 ref8">7, 8</xref>
        ].
      </p>
      <p>
        However, while NLP has demonstrated remarkable promise in addressing ofensive content,
its application to Low-Resource Indo-Aryan Languages presents a unique set of challenges.
Lowresource languages, such as Sinhala and Gujarati, often lack the extensive datasets and linguistic
resources available for major languages like English. Consequently, developing efective NLP
models for these languages becomes a complex endeavor, hindered by limited training data
and insuficient linguistic representation [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. Low-resource languages lack the necessary data
processing and ground truth data requirements for machine learning systems [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ].
      </p>
      <p>
        The urgency of mitigating ofensive content in Low-Resource Indo-Aryan Languages cannot
be understated. With millions of users communicating in these languages on various platforms,
there is a pressing need to safeguard them from the harmful impact of ofensive language. To
address this need, we propose a solution based on Multilingual Toxicity Analysis, a cutting-edge
approach that transcends language barriers and facilitates the detection of ofensive content
across multiple languages [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ].
      </p>
      <p>In this research, we embark on an experimental journey, aiming to harness the power of
transfer learning and zero-shot transfer techniques. Our goal is to create a model capable
of identifying hate speech and ofensive content in Hindi, Sinhala, Gujarati, Bengali, Bodo,
Assamese languages despite being primarily trained on English data through few-shot learning
with relatively smaller datasets. By strategically transferring knowledge from high-resource
languages like English to low-resource languages, we aim to enhance the capabilities of NLP
models in identifying toxic language across linguistic boundaries.</p>
      <p>The subsequent sections of this paper will delve into the methodology employed, including
data collection and preprocessing techniques, the selection and fine-tuning of NLP models, as
well as the evaluation metrics used to assess model performance. We will discuss the insights
drawn from our experiments, shedding light on the model’s adaptability in diverse linguistic
contexts. Additionally, we will explore the implications and limitations of our approach, paving
the way for further advancements in the realm of Multilingual Toxicity Analysis for
lowresource languages. Ultimately, our research strives to contribute to a safer and more inclusive
digital ecosystem by demonstrating that, with strategic fine-tuning, NLP models can transcend
language barriers and efectively detect ofensive content. By doing so, we aspire to create online
spaces where ofensive content is promptly identified and addressed, fostering an atmosphere
of respect and tolerance among users.</p>
    </sec>
    <sec id="sec-3">
      <title>2. Literature Review</title>
      <p>
        The literature surrounding hate speech and ofensive content detection has witnessed substantial
growth due to the increasing concern over the proliferation of harmful online behavior [
        <xref ref-type="bibr" rid="ref1 ref3">1, 3</xref>
        ].
This literature review aims to explore existing research in the field and shed light on innovative
approaches for addressing this pressing issue. By conducting a comprehensive analysis of prior
studies, we seek to identify new perspectives, reveal gaps in the literature, resolve conflicting
ifndings, and prevent duplication of efort.
      </p>
      <p>
        Several foundational studies have contributed significantly to the development of Natural
Language Processing (NLP) models as central characters in the fight against ofensive content
[
        <xref ref-type="bibr" rid="ref4">4, 12</xref>
        ]. DistilBERT [13] and XLM-RoBERTa [14] are two such prominent BERT [15] models that
have shown promise in detecting hate speech and ofensive language across diverse languages
and contexts. Researchers have devoted considerable efort to honing and refining these
characters, exploring their strengths and limitations in identifying toxic language patterns.
      </p>
      <p>
        The setting of this literature review encompasses the digital landscape, where social media
platforms and online communities serve as primary channels for communication. In this
virtual environment, ofensive content often thrives, necessitating efective content moderation
strategies [
        <xref ref-type="bibr" rid="ref4">16, 4</xref>
        ]. Understanding this setting is crucial for contextualizing the challenges and
opportunities for NLP models in detecting and mitigating hate speech and ofensive language
efectively.
      </p>
      <p>The plot of this literature review revolves around the development and evaluation of NLP
models tailored for hate speech and ofensive content detection. Researchers have embarked
on a journey to explore various embedding-classifier pairs, employing techniques such as
DistilBERT with Decision Tree, Gaussian Naive Bayes, Neural Network, and XLM-RoBERTa
embedder with its corresponding classifier [ 14, 17]. The plot thickens as researchers delve
into the intricacies of these models, seeking to improve performance and generalization across
diverse languages and contexts.</p>
      <p>
        The overarching theme that unifies the literature is the pursuit of fostering a safer online
environment and promoting linguistic inclusivity. By advancing multilingual toxicity
analysis, researchers contribute to creating digital ecosystems where users from diverse linguistic
backgrounds can engage in respectful and tolerant discourse [
        <xref ref-type="bibr" rid="ref2">18, 2</xref>
        ]. The theme emphasizes the
significance of breaking barriers, both linguistic and cultural, using advanced NLP techniques
and cross-lingual representation learning to build models that transcend language-specific
limitations.
      </p>
      <p>
        This literature review is framed within the context of a broader research landscape that
spans computational linguistics, NLP, and machine learning. It situates itself at the forefront of
addressing the challenges posed by ofensive content in Low-Resource Indo-Aryan Languages
like Sinhala and Gujarati [
        <xref ref-type="bibr" rid="ref9">9, 19, 20</xref>
        ]. By contextualizing our review within this frame, we
underscore the significance of our research in bridging gaps and ofering insights to the broader
academic community.
      </p>
      <p>The exposition in this literature review provides a comprehensive overview of the current
state of research in hate speech and ofensive content detection. We delve into existing studies,
methodologies, and NLP models, outlining the strengths and limitations of various approaches.
Through this exposition, we aim to provide a solid foundation for understanding the challenges
and potential solutions in multilingual toxicity analysis.</p>
      <p>In conclusion, this literature review serves as a critical link between the introduction and
methodology sections of our research. By exploring the foundational elements of NLP models
as characters, the digital setting, the plot of NLP techniques for ofensive content detection,
the overarching theme of fostering a safer online environment, and the framing within the
broader research landscape, we gain a deeper understanding of the landscape of multilingual
toxicity analysis. Our analysis aims to identify new avenues for research, resolve conflicts in
existing studies, and contribute to the collective efort of creating a safer and more inclusive
digital ecosystem.</p>
    </sec>
    <sec id="sec-4">
      <title>3. Methodology</title>
      <sec id="sec-4-1">
        <title>3.1. Data processing</title>
        <p>3.1.1. English
The dataset utilized was from the Kaggle jigsaw-toxic-comment-classification competition. It
had the following toxicity types:</p>
        <sec id="sec-4-1-1">
          <title>1. toxic</title>
          <p>2. severe_toxic</p>
        </sec>
        <sec id="sec-4-1-2">
          <title>3. obscene</title>
          <p>4. threat</p>
        </sec>
        <sec id="sec-4-1-3">
          <title>5. insult</title>
          <p>6. identity_hate</p>
          <p>Since they are all regarded as harmful, if a comment contained at least one of the above labels,
we regarded it as toxic and combined the comments into a new super column. The dataset
approximately had 150k samples of labelled data, after dropping the null values. Moreover, the
data was cleaned by removing unnecessary symbols, tags, etc. For training 80% of the samples
were used while 20% were stored for validation and evaluation.
3.1.2. Hindi
The hindi dataset from [21] was used for hindi fine-tuning. The dataset size had 1603 training
samples in text format with labels. The types of unique labels that were found after being
separated from text included the following:
1. __label__abusive,
2. __label__abusive__label__hatespeech,
3.
__label__abusive__label__hate</p>
          <p>speech__label__humor,
4. __label__abusive__label__humor,
5. __label__benign,
6. __label__benign__label__abusive,
7. __label__benign__label__humor,
8. __label__hatespeech,
9. __label__hatespeech__label__abusive,
10.
__label__hatespeech__label__abu</p>
          <p>sive__label__abusive,
11. __label__hatespeech__label__benign,
12. __label__humor__label__abusive,
13. __label__humor__label__benign,
14.
__label__humor__label__hatespeech__label__abusive</p>
          <p>We labelled the ones containing ’label__abusive’ or ’label__hatespeech’ as ’HOF’ and the rest
as ’NOT’ where ’HOF’ is encoded further as 1, and ’NOT’ as 0. The train set had 1603 samples
while test had 397.
3.1.3. Gujarati
The Gujarati dataset contains 200 tweets. Firstly [22, 23], the tweets were cleaned by removing
unnecessary symbols, tags, and mentions. Moreover, the labels ’HOF’ and ’NOT’ in the dataset
were encoded to 1 and 0 respectively. The same procedure was used for test set which contained
1196 samples.
3.1.4. Bengali, Assamese, Sinhala and Bodo
The datasets of Bengali [24, 23], Assamese [25, 24, 23], Sinhala [26, 23], and Bodo [24, 23] are
primarily collected from Twitter, Facebook, or youtube comments, all were cleaned with the
same process as Gujarati, and had the same label convention.
3.1.5. Findings from datasets
After analyzing each dataset, it was found that each sample had, on average, fewer than 100
tokens. Based on these findings we decided the ’max_length’ of the model. Moreover, for
tokenization ’word_tokenize’ from ’nltk’ was used to tokenize the samples, and average tokens
(avg_tokens) were calculated using the average formula that is:
 
_ =</p>
          <p>∑=1 ∑=1</p>
          <p>∑=1 
where:</p>
        </sec>
        <sec id="sec-4-1-4">
          <title>1. n: total number of samples</title>
          <p>2. T : token of a word</p>
        </sec>
      </sec>
      <sec id="sec-4-2">
        <title>3.2. Model and architecture</title>
        <sec id="sec-4-2-1">
          <title>3. m: total words in a sample</title>
        </sec>
        <sec id="sec-4-2-2">
          <title>4. l: total number of samples</title>
          <p>We employ the XLM-RoBERTa model, a powerful pre-trained language representation model, for
the task of hate speech and ofensive content detection. XLM-RoBERTa, short for Cross-lingual
Language Model - RoBERTa, is an extension of the RoBERTa model that is specifically designed
to handle multilingual text, and performs particularly well on resource languages[14].</p>
          <p>The architecture of XLM-RoBERTa consists of an embeddings layer, followed by an encoder
layer. The embeddings layer includes word embeddings, position embeddings, and token type
embeddings, which together form the input representation for the model. The encoder layer
comprises a series of XLM-RoBERTa layers, each consisting of attention mechanisms and
feed-forward neural networks [14].</p>
          <p>The attention mechanism in XLM-RoBERTa enables the model to focus on important parts
of the input sequence during processing. This mechanism includes self-attention, where each
word in the input is associated with weights that determine its relevance to other words in the
sequence [27].</p>
          <p>Furthermore, XLM-RoBERTa incorporates LayerNorm and dropout layers to improve model
stability and prevent overfitting. LayerNorm normalizes the hidden state of each layer, and
dropout randomly deactivates certain neurons during training to prevent the model from relying
too heavily on specific features [ 28, 29].</p>
          <p>The classification head of XLM-RoBERTa consists of a dense layer, followed by a dropout layer,
and an output projection layer. The dense layer helps in reducing the feature dimensionality,
while dropout aids in regularization to avoid overfitting. The output projection layer maps the
reduced features to the final output, which is a binary classification for hate speech and ofensive
content detection [14]. The model size is 278M parameters. However, trainable classification
head has 592,130 parameters. Additionally, the maximum length of the model was set at 256
for Code-Mixed Languages and 128 for the rest based on the results of the dataset processing
because the average token count was less than 100. The figure 1 below illustrates sample model
pipeline. The pipeline consist of an multi-lingual embedder followed by a classifier:</p>
          <p>In summary, the XLM-RoBERTa model, with its multilingual representation learning
capabilities and attention-based architecture, is a robust choice for our hate speech and ofensive
content detection task. By leveraging its pre-trained language understanding, we aim to achieve
accurate and efective results across diverse languages and contexts.</p>
        </sec>
      </sec>
      <sec id="sec-4-3">
        <title>3.3. Model Implementation</title>
        <p>We propose a model (the implementation can be viewed at https://github.com/Jayveersinh-Raj/
indo_aryan_clf_HASOC_2023) which is initially trained on an English dataset and subsequently
ifne-tuned on diferent languages. By following the above approach the model can be tailored
to any low resource language with relatively small dataset. The naming conventions for the
models are as follows:
• XLM-Classifier : The radomly initialized classifier not finetuned for downstream task.
• PolyGuard: The English only trained model (XLM-Classifier) [ 30]
• Indo-Aryan-Extension: The PolyGuard finetuned on Hindi [ 31]
• xlm-sinhala: The Sinhala only trained model (XLM-Classifier)
• xlm-bengali: The Bengali only trained model (XLM-Classifier)
• xlm-bodo: The Bodo only trained model (XLM-Classifier)
• xlm-assamese: The Assamese only trained model (XLM-Classifier)
It is important to note that in all of the models above, and if fine-tuned, we utilized the AdamW
optimizer which is an extension of the Adam optimizer using a method known as weight decay.
To stop overfitting, a regularization factor called weight decay is applied to the loss function.
By separating weight decay from the optimization processes, AdamW solves the weight decay
problems with the original Adam optimizer, providing more reliable and eficient training. The
common hyperparameters used are as follows:
• learning rate: 210 −5
• epsilon: 110 −8
• Trainable Layers: All
These hyper-parameters were chosen to balance precision in weight updates and numerical
stability during the optimization process, ensuring the model’s efective training and
performance. Another important hyperparameter we used was epoch, which is mentioned in each of
the sub-experiments in the results section. We tested diferent epochs on each sub-dataset to
ifnd the best performance, and we chose the epochs that performed well on the unseen test
data, which helped to mitigate overfitting and underfitting.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>4. Results</title>
      <p>In this section, we present the results of our experiments in fine-tuning the XLM-RoBERTa
model for hate speech and ofensive content detection in various Indian languages. The
experiments considers fixing the model while testing the performance on various datasets, and few
shot capabilities of the model. The primary focus for few shot inference on languages such
as Gujarati which has relatively low size leaving no room to train a raw architecture on the
dataset.</p>
      <p>We report the dataset details and performance metrics for each language below, however, it
is imperative to mention that in order to avoid over-fitting we verified the outcomes by training
the models for a slightly shorter number of epochs than stated below for each language.</p>
      <sec id="sec-5-1">
        <title>4.1. English</title>
        <p>Dataset Details:</p>
        <sec id="sec-5-1-1">
          <title>Model</title>
          <p>XLM-Classifier
• Training Dataset: Approximately 150k samples encoded as 1 for Hatespeech, and 0 for
neutral
• Total Training Samples: Approximately 120k
• Test Dataset: Approximately 30k.
epochs F1-score
3 0.960
• Training Dataset: Cleaned with sentences split from labels. ’Hatespeech’ labeled as ’HOF’
(1), the rest as ’NOT’ (0).
• Total Training Samples: 1603
• Test Dataset: Similar preprocessing applied, containing 397 samples.
4.2. Hindi</p>
        </sec>
      </sec>
      <sec id="sec-5-2">
        <title>4.3. Gujarati</title>
        <p>• Training Dataset: Contains 200 crawled tweets in Gujarati labeled ’HOF’ (1) for hate
speech and ’NOT’ (0) for neutral.
• Unique local slangs observed in tweets, cleaned by removing mentions and symbols.
• Test set contained 1196 tweets</p>
        <p>Model performances
Model epochs F1-score
Direct inference (PolyGuard) - 0.165
Direct inference (Indo-Aryan- - 0.547
Extension)
Fine-tuned (PolyGuard) 20 0.791
Fine-tuned (Indo-Aryan- 20 0.793
Extension)</p>
        <sec id="sec-5-2-1">
          <title>Leaderboard Rank</title>
          <p>4</p>
        </sec>
      </sec>
      <sec id="sec-5-3">
        <title>4.4. Sinhala</title>
        <p>Dataset Details:
• Training Dataset: Contains 7,500 samples in Sinhala labeled ’HOF’ (1) for hate speech
and ’NOT’ (0) for neutral.
• Test Samples: 2,500.</p>
        <sec id="sec-5-3-1">
          <title>Model</title>
          <p>xlm-sinhala
Fine-tuned
Extension)
Fine-tuned (PolyGuard)
4.6. Bodo</p>
        </sec>
      </sec>
      <sec id="sec-5-4">
        <title>4.7. Assamese</title>
        <p>Dataset Details:</p>
      </sec>
      <sec id="sec-5-5">
        <title>4.5. Bengali</title>
      </sec>
      <sec id="sec-5-6">
        <title>4.8. General Analysis</title>
        <p>• The fine-tuning process demonstrates the model’s adaptability to diferent languages
with significantly low fine tuning samples of the target language, with F1 scores ranging
from 0.70 to 0.89.
• Notably, the model performs exceptionally well in languages like Sinhala and Bodo, where</p>
        <p>F1 scores exceed 0.8.
• The performance in Assamese and Bengali datasets is moderate, suggesting room for
further improvement.</p>
        <p>These experiments demonstrate the XLM-RoBERTa model’s potential for cross-lingual hate
speech detection, as well as the potential of our ’PolyGuard’ model, which, after being trained
on 100k+ samples of English, can produce significant results after being fine-tuned with a very
small dataset of a target language. The results vary across languages, emphasizing the need for
language-specific adaptations and larger, diverse datasets to enhance model performance further.
Future work should focus on addressing the challenges posed by low resource languages and
exploring techniques to improve classification accuracy.</p>
        <p>The following figure demonstrates that fine tuned ’PolyGuard’ yields better results over multiple
languages:
Moreover, the results of the fine-tuned ’PolyGuard’ model on varying dataset size can be
seen in the figure below:</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>5. Discussion</title>
      <p>In this section, we delve into the insights drawn from the results obtained in our experiments.
The varying performance across languages and datasets provides valuable insights into the
challenges and opportunities in multilingual hate speech detection.</p>
      <sec id="sec-6-1">
        <title>5.1. Cross-Lingual Adaptability</title>
        <p>Our experiments have showcased the remarkable cross-lingual adaptability of the XLM-RoBERTa
model. It excelled in languages like Sinhala and Bodo, where the F1 scores exceeded 0.8. This
demonstrates the model’s potential for efectively identifying hate speech in languages with
relatively larger datasets and well-defined linguistic patterns. It aligns with prior research
indicating that multilingual models can leverage their pre-trained knowledge efectively across
languages [32].</p>
      </sec>
      <sec id="sec-6-2">
        <title>5.2. Data Quality and Size</title>
        <p>The performance in Gujarati is notable, with an F1 score of 0.7926. It’s essential to highlight
that this dataset contained local slangs and expressions that might only be discernible to native
speakers. This raises questions about data quality and the importance of domain-specific
knowledge in hate speech detection tasks. Additionally, the dataset sizes played a crucial role in
model performance. Languages with larger datasets, such as Sinhala and Bodo, yielded higher
F1 scores. Hence, expanding and diversifying datasets for low-resource languages is essential
for improved model performance.</p>
      </sec>
      <sec id="sec-6-3">
        <title>5.3. Room for Improvement</title>
        <p>While our model achieved competitive results, there is room for improvement. For instance, in
Bengali and Assamese datasets, where F1 scores were moderate (0.726 and 0.707 respectively),
further fine-tuning, data augmentation, or specialized techniques for handling local dialects may
enhance performance. Additionally, exploring more advanced model compression techniques
could optimize eficiency while retaining performance [ 33].</p>
      </sec>
      <sec id="sec-6-4">
        <title>5.4. Future Directions</title>
        <p>Future work in cross-lingual hate speech detection should focus on:
1. Investigating techniques for better handling dialects.
2. Expanding and diversifying datasets for low-resource languages.
3. Exploring advanced fine-tuning strategies to boost model performance.
4. Addressing issues related to model eficiency and interpretability.</p>
        <p>Our experiments underscore the adaptability and potential of the XLM-RoBERTa model in
multilingual hate speech detection. However, they also shed light on the complexities and the
critical role of dataset size and quality. These findings provide a foundation for further research
aimed at bridging the gap in hate speech detection across diverse languages and cultures.</p>
      </sec>
    </sec>
    <sec id="sec-7">
      <title>6. Conclusion</title>
      <p>In this study, we embarked on a journey to break barriers in multilingual toxicity analysis,
focusing on hate speech and ofensive content detection in low-resource languages. Leveraging
the formidable capabilities of the XLM-RoBERTa model, we conducted a series of experiments
ifne-tuning the model for various languages. The insights drawn from our endeavors provide
valuable contributions to the field of cross-lingual hate speech detection.</p>
      <p>Our results are promising, indicating the remarkable cross-lingual adaptability of the
XLMRoBERTa model. It demonstrated exceptional performance in languages such as Sinhala and
Bodo, achieving F1 scores exceeding 0.8. These outcomes underscore the potential of
multilingual models to efectively identify hate speech in languages with relatively larger datasets
and well-defined linguistic patterns. This aligns with the vision of creating models that can
transcend language barriers.</p>
      <p>Our experiments also emphasized the critical importance of data quality and dataset size. In
Gujarati, the model achieved commendable results (F1 score of 0.7926), but the dataset contained
local slangs and expressions that might be understood only by native speakers. This raises
questions about the significance of domain-specific knowledge in hate speech detection tasks.
Additionally, languages with larger datasets, like Sinhala and Bodo, yielded higher F1 scores,
highlighting the importance of dataset expansion and diversification for low-resource languages.</p>
      <p>Despite these challenges, our model has exhibited its potential to deliver competitive
performance in multilingual hate speech detection. With access to balanced datasets and further
ifne-tuning eforts, we believe our model can surpass competitors in the field, making significant
strides in mitigating hate speech and ofensive content in diverse linguistic landscapes. This
study has also demonstrated the generalizability of our model to diferent datasets, paving the
way for its broader application.</p>
      <p>In conclusion, our journey in breaking barriers through multilingual toxicity analysis has
yielded promising results. The XLM-RoBERTa model’s adaptability, coupled with the insights
gained from our experiments, sets the stage for future research and advancements in hate
speech detection across languages and cultures. As we continue to refine and expand our model,
we are one step closer to fostering more inclusive and respectful online spaces for all.
[12] W. Liu, S. Singh, M. Gardner, P. Kohli, Deep learning for extreme multi-label text
classification, arXiv preprint arXiv:1905.07342 (2019).
[13] V. Sanh, L. Debut, J. Chaumond, T. Wolf, Distilbert, a distilled version of bert: smaller,
faster, cheaper and lighter, 2020. arXiv:1910.01108.
[14] A. Conneau, G. Lample, Unsupervised cross-lingual representation learning at scale, in:
Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics,
2019, pp. 8440–8451.
[15] K. Ghosh, A. Senapati, U. Garain, Baseline bert models for conversational hate speech
detection in code-mixed tweets utilizing data augmentation and ofensive language identification
in marathi, in: Fire, 2022. URL: https://api.semanticscholar.org/CorpusID:259123570.
[16] I. Kwok, Y. Wang, Locate the hate: Detecting tweets against blacks, Proceedings of the
51st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long
Papers) (2013) 1250–1259.
[17] J. Devlin, M.-W. Chang, K. Lee, K. Toutanova, Bert: Pre-training of deep bidirectional
transformers for language understanding, 2019. arXiv:1810.04805.
[18] W. Yin, A. Zubiaga, Towards generalisable hate speech detection: a review on obstacles
and solutions, 2021. arXiv:2102.08886.
[19] A. Vadesara, P. Tanna, Corpus building for hate speech detection of gujarati language, in:
International Conference on Soft Computing and its Engineering Applications, Springer,
2022, pp. 382–395.
[20] H. Sandaruwan, S. Lorensuhewa, M. Kalyani, Sinhala hate speech detection in social
media using text mining and machine learning, in: 2019 19th International Conference on
Advances in ICT for Emerging Regions (ICTer), volume 250, IEEE, 2019, pp. 1–8.
[21] V. K. Jha, H. P, V. P. N, V. Vijayan, P. P, Dhot-repository and classification of ofensive
tweets in the hindi language, Procedia Computer Science 171 (2020) 2324–2333. doi:https:
//doi.org/10.1016/j.procs.2020.04.252.
[22] S. Satapara, H. Madhu, T. Ranasinghe, A. E. Dmonte, M. Zampieri, P. Pandya, N. Shah,
M. Sandip, P. Majumder, T. Mandl, Overview of the hasoc subtrack at fire 2023:
Hatespeech identification in sinhala and gujarati, in: K. Ghosh, T. Mandl, P. Majumder, M. Mitra
(Eds.), Working Notes of FIRE 2023 - Forum for Information Retrieval Evaluation, Goa,
India. December 15-18, 2023, CEUR Workshop Proceedings, CEUR-WS.org, 2023.
[23] T. Ranasinghe, K. Ghosh, A. S. Pal, A. Senapati, A. E. Dmonte, M. Zampieri, S. Modha,
S. Satapara, Overview of the HASOC subtracks at FIRE 2023: Hate speech and ofensive
content identification in assamese, bengali, bodo, gujarati and sinhala, in: Proceedings of
the 15th Annual Meeting of the Forum for Information Retrieval Evaluation, FIRE 2023,
Goa, India. December 15-18, 2023, ACM, 2023.
[24] K. Ghosh, A. Senapati, A. S. Pal, Annihilate Hates (Task 4, HASOC 2023): Hate Speech
Detection in Assamese, Bengali, and Bodo languages, in: Working Notes of FIRE 2023
Forum for Information Retrieval Evaluation, CEUR, 2023.
[25] K. Ghosh, D. Sonowal, A. Basumatary, B. Gogoi, A. Senapati, Transformer-based hate
speech detection in assamese, in: 2023 IEEE Guwahati Subsection Conference (GCON),
2023, pp. 1–5. doi:10.1109/GCON58516.2023.10183497.
[26] T. Ranasinghe, I. Anuradha, D. Premasiri, K. Silva, H. Hettiarachchi, L. Uyangodage,</p>
      <p>M. Zampieri, Sold: Sinhala ofensive language dataset, 2022. arXiv:2212.00851.
[27] A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, Ł. Kaiser, I.
Polosukhin, Attention is all you need, Advances in Neural Information Processing Systems 30
(2017) 5998–6008.
[28] J. L. Ba, J. R. Kiros, G. E. Hinton, Layer normalization, arXiv preprint arXiv:1607.06450
(2016).
[29] N. Srivastava, G. Hinton, A. Krizhevsky, I. Sutskever, R. Salakhutdinov, Dropout: a simple
way to prevent neural networks from overfitting, Journal of Machine Learning Research
15 (2014) 1929–1958.
[30] J. Raj, Polyguard model, 2023. URL: https://huggingface.co/Jayveersinh-Raj/PolyGuard.
[31] J. Raj, Indo-aryan-abuse-detection, 2023. URL: https://huggingface.co/Jayveersinh-Raj/</p>
      <p>Indo-Aryan-abuse-detection.
[32] Y. Tang, C. Tran, X. Li, P.-J. Chen, N. Goyal, V. Chaudhary, J. Gu, A. Fan, Multilingual
translation with extensible multilingual pretraining and finetuning, 2020. arXiv:2008.00401.
[33] A. Berthelier, T. Chateau, S. Dufner, C. Garcia, C. Blanc, Deep model compression and
architecture optimization for embedded systems: A survey, Journal of Signal Processing
Systems 93 (2021) 863–878.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>X.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Luo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Hu</surname>
          </string-name>
          ,
          <article-title>Detecting hate speech on twitter using a convolution-gru based deep neural network</article-title>
          ,
          <source>Proceedings of the 27th international conference on computational linguistics</source>
          (
          <year>2018</year>
          )
          <fpage>2561</fpage>
          -
          <lpage>2571</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>T.</given-names>
            <surname>Davidson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Warmsley</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Macy</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Weber</surname>
          </string-name>
          ,
          <article-title>Automated hate speech detection and the problem of ofensive language</article-title>
          ,
          <source>Proceedings of the 11th international conference on web and social media</source>
          (
          <year>2017</year>
          )
          <fpage>512</fpage>
          -
          <lpage>515</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Waseem</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Hovy</surname>
          </string-name>
          ,
          <article-title>Hateful symbols or hateful people? predictive features for hate speech detection on twitter, NAACL-HLT (</article-title>
          <year>2016</year>
          )
          <fpage>88</fpage>
          -
          <lpage>93</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>A.</given-names>
            <surname>Schmidt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Wiegand</surname>
          </string-name>
          ,
          <article-title>A survey of hate speech detection using natural language processing</article-title>
          ,
          <source>Proceedings of the fith international workshop on natural language processing for social media</source>
          (
          <year>2017</year>
          )
          <fpage>1</fpage>
          -
          <lpage>10</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>C. J.</given-names>
            <surname>Hutto</surname>
          </string-name>
          , E. Gilbert,
          <article-title>Vader: A parsimonious rule-based model for sentiment analysis of social media text, Eighth international conference on weblogs and social media (</article-title>
          <year>2014</year>
          )
          <fpage>216</fpage>
          -
          <lpage>225</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>A.</given-names>
            <surname>Go</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Bhayani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <article-title>Twitter sentiment classification using distant supervision</article-title>
          ,
          <source>CS224N Project Report, Stanford</source>
          <volume>1</volume>
          (
          <year>2009</year>
          )
          <year>2009</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Waseem</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Hovy</surname>
          </string-name>
          ,
          <article-title>Understanding abuse: A typology of abusive language detection subtasks</article-title>
          ,
          <source>Proceedings of the first workshop on abusive language online</source>
          (
          <year>2017</year>
          )
          <fpage>1</fpage>
          -
          <lpage>10</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>K.</given-names>
            <surname>Ghosh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. A.</given-names>
            <surname>Senapati</surname>
          </string-name>
          ,
          <article-title>Hate speech detection: a comparison of mono and multilingual transformer model with cross-language evaluation</article-title>
          ,
          <source>in: Proceedings of the 36th Pacific Asia Conference on Language, Information and Computation</source>
          , De La Salle University, Manila, Philippines,
          <year>2022</year>
          , pp.
          <fpage>853</fpage>
          -
          <lpage>865</lpage>
          . URL: https://aclanthology.org/
          <year>2022</year>
          .paclic-
          <volume>1</volume>
          .
          <fpage>94</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>M.</given-names>
            <surname>Geva</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Goldberg</surname>
          </string-name>
          ,
          <article-title>Contextualizing hate speech classifiers with post-hoc explanations</article-title>
          ,
          <source>Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLPIJCNLP)</source>
          (
          <year>2019</year>
          )
          <fpage>2736</fpage>
          -
          <lpage>2741</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>D.</given-names>
            <surname>Nkemelu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Shah</surname>
          </string-name>
          , I. Essa,
          <string-name>
            <given-names>M. L.</given-names>
            <surname>Best</surname>
          </string-name>
          ,
          <article-title>Tackling hate speech in low-resource languages with context experts</article-title>
          ,
          <year>2023</year>
          . arXiv:
          <volume>2303</volume>
          .
          <fpage>16828</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>T.</given-names>
            <surname>Schuster</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Bastings</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Belinkov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Goldwater</surname>
          </string-name>
          ,
          <article-title>Cross-lingual alignment of contextual word embeddings, with applications to zero-shot dependency parsing</article-title>
          ,
          <source>in: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics</source>
          ,
          <year>2019</year>
          , pp.
          <fpage>3500</fpage>
          -
          <lpage>3508</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>