<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>Forum for Information Retrieval Evaluation, December</journal-title>
      </journal-title-group>
      <issn pub-type="ppub">1613-0073</issn>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>BERT-Based Analysis of Bengali, Assamese, &amp; Bodo Conversational Hateful Content from Social Media</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Jhuma Kabir Mim</string-name>
          <email>jhuma.mim@student.lut.fi</email>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Mourad Oussalah</string-name>
          <email>mourad.oussalah@oulu.fi</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Akash Singhal</string-name>
          <email>akash.singhal@helsinki.fi</email>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>CMVS, Faculty of ITEE, University of Oulu</institution>
          ,
          <addr-line>Oulu</addr-line>
          ,
          <country country="FI">Finland</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>CVPR , LUT University</institution>
          ,
          <addr-line>Lappeenranta</addr-line>
          ,
          <country country="FI">Finland</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Dept of Computer Science, University of Helsinki</institution>
          ,
          <addr-line>Helsinki</addr-line>
          ,
          <country country="FI">Finland</country>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>used BERT models</institution>
          ,
          <addr-line>including XML-Roberta, L3-cube, IndicBERT, BenglaBERT, and BanglaHateBERT</addr-line>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2023</year>
      </pub-date>
      <volume>1</volume>
      <fpage>5</fpage>
      <lpage>18</lpage>
      <abstract>
        <p>In today's age, social media reigns as the paramount communication platform, providing individuals with the avenue to express their conjectures, intellectual propositions, and reflections. Unfortunately, this freedom often comes with a downside as it facilitates the widespread proliferation of hate speech and ofensive content, leaving a deleterious impact on our world. Thus, it becomes essential to discern and eradicate such ofensive material from the realm of social media. This article delves into the comprehensive results and key revelations from the HASOC-2023 ofensive language identification result. The primary emphasis is placed on the meticulous detection of hate speech within the linguistic domains of Bengali, Assamese, and Bodo, forming the framework for Task 4: Annihilate Hates. In this work, we The research outcomes were promising and showed that XML-Roberta-lagre performed better than monolingual models in most cases. Our team 'TeamBD' achieved rank 3rd for Task 4 - Assamese, &amp; 5th ∗Corresponding author. †These authors contributed equally. (A. Singhal) htp:/ceur-ws.org CEUR Workshop Proceedings (CEUR-WS.org) ISN1613-073</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>CEUR
ceur-ws.org</p>
    </sec>
    <sec id="sec-2">
      <title>1. Introduction</title>
      <p>
        Social media platforms such as Facebook, Twitter, and YouTube stand out as a widely embraced
and efortless avenue for uninhibited self-expression and online interaction. Regrettably, it
also serves as a platform for disseminating harmful and hostile content, including gender
discrimination, xenophobia, protests over politics, online harassment, and even extortion [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ].
Hate speech [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] involves derogatory comments targeting an individual or a group due to certain
attributes like race, color, or ethnicity. Now this profane language is pervasive on social media, it
has become a challenge. Researchers and organizations have been working to develop techniques
that can recognize hate speech or abusive language and flag it for review by human moderators
or for automated elimination. Consequently, numerous social media platforms actively monitor
CEUR
Workshop
Proceedings
user posts. This includes promoting the development of automated techniques for detecting or at
least provide insights to detect suspicious and harmful posts. Previous research has explored the
identification of ofensive language on various platforms, including Twitter [
        <xref ref-type="bibr" rid="ref1 ref3 ref4">3, 4, 1</xref>
        ], Wikipedia
comments, and Facebook posts [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], FromSpring posts [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], YouTube [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], and news articles
[
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], mostly conducted in the English language. The main obstacles in hate speech detection
revolve around the lack of essential resources with specific language datasets. Languages with
limited resources [
        <xref ref-type="bibr" rid="ref10 ref9">9, 10</xref>
        ] encounter the dificulty of having inadequately annotated datasets,
with only a restricted availability of monolingual datasets. Taking this situation into account,
Task 4 of HASOC-23 has the objective of detecting hate speech in Bengali, Bodo, and Assamese
languages in 2023 shared task challenge. Advancements [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] in natural language processing
(NLP) technology have spurred extensive research into the automated detection of textual hate
speech in recent years. Numerous studies [
        <xref ref-type="bibr" rid="ref12">12, 13</xref>
        ] have centered their eforts on employing
deep learning-based models for the detection of aggressive language within social media text.
Academics have also employed FastText embeddings to construct models capable of being
trained on billions of words in under ten minutes and classifying millions of sentences into
hundreds of categories [14]. There have also been studies [15] indicating how the bias and
worldview of annotators can impact the performance of such datasets. Currently, the
state-ofthe-art research in hate speech detection has advanced to the stage where researchers harness
the capabilities of advanced architectures like transfer learning. For instance, in [16], researchers
conducted a comparative analysis of deep learning, transfer learning, and multitask learning
architectures using an Arabic hate speech dataset. Additionally, another research focused on
identifying hate speech within Bengali and hindi datasets through the use of state-of-the-art
pretrained models such as XML-Roberta and multilingual BERT(mBERT) [17]. One major
hurdle lies in the fact that Bengali is considered a low-resource language. Regrettably, there is a
limited amount of research conducted on hate speech detection within Bengali social media
resources. Extensive research eforts have been dedicated to developing word embeddings
tailored specifically for low-resource languages, with an example being BengFastText [ 18],
designed for Bengali language. In addition, a studied carried out in [19] showed the promise of
BanglaHateBERT -a retrained BERT model designed for detecting abusive language in Bengali
language. This model underwent training using a substantial corpus of ofensive, abusive, and
hateful content in Bengali collected from various sources. Likewise, authors of [20] compared
various various deep learning models along with pretrained Bengali word embeddings such
as Word2Vec, FastText, and BengFastText. Nevertheless, there has been limited research efort
conducted in the Assamese and Bodo languages due to low resource. Preliminary works were
carried out using transformer architecture (BanglaBERT and mBERT) for abusive language
detection in Assamese text [21] &amp; [22], although the authors acknowledged the limited scale and
scope of the underlined study, calling for further research eforts. NLP community organized
several initiatives to handle this challenge and stimulate research on hateful speech and ofensive
content in social media, such as Semeval-2019 [23] and 2020 [24]. In the HASOC-2020 [25]
&amp; 2021 [
        <xref ref-type="bibr" rid="ref13">26</xref>
        ] shared task, it is noteworthy that it drew participation from an extensive pool
of over 40 research groups. The attainment of the highest ranking in Hindi hate speech
detection was accomplished through the utilization of a Convolutional Neural Network (CNN)
incorporating FastText [
        <xref ref-type="bibr" rid="ref14">27</xref>
        ] embeddings as input. For the task of German hate speech detection,
superior performance was achieved through the deployment of fine-tuned iterations of BERT,
DistilBERT, and RoBERTa [
        <xref ref-type="bibr" rid="ref15">28</xref>
        ]. Similarly, the performance in English-language hate speech
detection was reached by leveraging BERT alongside another deep learning-based model. These
outcomes underscore the diverse and innovative approaches employed by diferent research
groups in addressing the challenge of hate speech detection in varying linguistic contexts.
Based on prior research, the BERT model has consistently surpassed the performance of many
other contemporary state-of-the-art models. The increasing prominence of BERT stands as a
significant trend, underscoring its popularity within the hate speech detection community. In
the last five years, BERT has accounted for a substantial share (38%) of deep learning models
employed for this purpose [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ].
      </p>
      <p>
        In this year for the first time, 2023 HASOC [
        <xref ref-type="bibr" rid="ref16 ref17 ref18">29, 30, 31</xref>
        ] introduced a novel task involving the
identification of multilingual ofensive languages for social media, encompassing platforms like
Facebook, Twitter, &amp; YouTube. In this contest, our team participated to Task 4 for Bengali, Bodo,
and Assamese languages identification of Hate or ofensive. In the context of the state-of-art
advancements in hate-speech or ofensive text detection, our paper contributes in the following
ways:
1. We compare Pre-trained BERT models such as XML-Roberta, L3-cube, IndicBERT,
      </p>
      <p>BanglaBERT, and BanglaHateBert for hate speech detection in low resource language.
2. We employed data augmentation using ChatGPT3.
3. We expanded the Dataset through Self-Model Annotation.</p>
      <p>The remaining sections of this paper are organized as follows. Section 2 explores the task
description and dataset for the three languages. Section 3 details our methodology, encompassing
details about the employed transformer models including feature engineering and presents the
experimental results, highlighting and discussing the best scoring model. Lastly, conclusive
statements and potential future work are drawn in the conclusion section.</p>
    </sec>
    <sec id="sec-3">
      <title>2. Task Description</title>
      <p>
        Task 4 focuses on identifying hate speech, ofensive language, and profanity in diferent
languages using natural language processing techniques. The objective of the task is to detect
hate speech in Bengali, Bodo, and Assamese languages. To accomplish this, we used the
HASOC2023 shared dataset for training and validation processes, and cast the problem into a binary
classification task [
        <xref ref-type="bibr" rid="ref16 ref19 ref20">29, 32, 33, 22</xref>
        ]. Each dataset (for the three languages) consists of a list of
sentences with their corresponding class (hate/ofensive (HOF) or non hate/ofensive (NOT)).
      </p>
      <sec id="sec-3-1">
        <title>2.1. Datasets</title>
        <p>Data is primarily collected from Twitter, Facebook and YouTube comments. The total size of
the three datasets all together amounts to 6996 comments, among which 3860 (80%) contain
HOF and the rest 3136 (20%) NOT.</p>
        <p>1. (NOT) Not-Hate or Ofensive - This post does not contain any Hate speech, profane,
ofensive content.</p>
        <p>2. (HOF) Hate or Ofensive - This post contains Hate, ofensive, and profane content.
"তেব শুনলাম মমতা ব্যানার্িজ েকাটা পদ্ধিত তু েল
িদেয়েছ তাহেল ব্রাত্য বসুর মুেখ েকাটা েকন?
এটা কথার কথা? তাহেল েকাটা িনেয় প্রাক্তন
মুখ্যমন্ত্রীেক কটাক্ষ েকন?"
"দুজেনই মাতাল মাগীবাজ"
"কু কু ৰ বুিল িকয় ৈকেছ অসভ্য ক'ৰবাৰ, লাজ নাই"
"मोसौ खगुायाव एमफौ नांबाय नोनंाव समै"</p>
        <sec id="sec-3-1-1">
          <title>Translation</title>
          <p>Stupid Education Minister</p>
        </sec>
        <sec id="sec-3-1-2">
          <title>Label</title>
          <p>HOF</p>
        </sec>
        <sec id="sec-3-1-3">
          <title>Language</title>
          <p>Bengali
However, I heard that
Mamata Banerjee has given
up the quota system, so why
is there a quota in Bratya
Bose’s mouth? Is it word of
mouth? So why sneer at the
former chief minister about
the quota?
Both are drunkards fuckers
Why are you calling me a
dog, rude somewhere, no
shame
NONE</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>3. Methodology</title>
      <p>This section provides a thorough overview that includes both the model architectures and
methodologies used to tackle the task.</p>
      <sec id="sec-4-1">
        <title>3.1. Models Descriptions</title>
        <p>
          BERT – short for Bidirectional Encoder Representations from Transformers, represents a
pioneering language model rooted in transformer architecture [
          <xref ref-type="bibr" rid="ref21">34</xref>
          ]. This influential model
employs an attention mechanism, thereby enabling the acquisition of contextual relationships
among words in a given text sequence. BERT adopts two primary training strategies:
1. Masked Language Modeling (MLM) where approximately 15% of the tokens within a
sequence are replaced (masked), prompting the model to predict the original tokens.
2. Next Sentence Prediction (NSP) where the model is presented with two sentences as input,
and it learns to determine whether the second sentence follows the first in their original
document context.
        </p>
        <p>
          XML-Roberta – XML-Roberta is a variant of the RoBERTa model, which is a popular
transformer-based language model because of its ability to handle cross-lingual tasks. It has
been fine-tuned to perform exceptionally well in various languages and is particularly robust
in multilingual settings. The ”XML” refers to ”Cross-lingual Multilingual,” emphasizing its
proficiency in understanding and generating text across more than 100 languages [
          <xref ref-type="bibr" rid="ref22">35</xref>
          ] at once.
        </p>
        <p>
          IndicBERT – IndicBERT represents a multilingual ALBERT model that has been exclusively
pretrained on a comprehensive dataset comprising 12 major Indian languages [
          <xref ref-type="bibr" rid="ref23">36</xref>
          ]. Following
pretraining, IndicBERT underwent evaluation across a range of diverse natural language
understanding tasks.
        </p>
        <p>
          L3-cube – It is a mBERT model fine-tuned on L3Cube-HingCorpus. The latter is the first
large-scale real Hindi-English code mixed data known as HingBERT [
          <xref ref-type="bibr" rid="ref24">37</xref>
          ] in a Roman script.
        </p>
        <p>BanglaBERT – Bangla-Bert-Base [38] is a pre-trained Bengali language model that employs
the mask language modeling technique, as outlined in the BERT framework.</p>
        <p>BanglaHateBERT– BanglaHateBERT [19] model is essentially designed for abusive and
ofensive language identification in Bengali.</p>
      </sec>
      <sec id="sec-4-2">
        <title>3.2. Experiments and Results</title>
        <p>In the context of Task 4, we embraced three primary steps- based methodology:</p>
        <p>Exploration stage: We conducted experiments with a range of BERT models, both
multilingual (e.g., XML-Roberta-large and IndicBERT) and monolingual (e.g., L3-cube, Bangla
BERT, and Bangla Hate BERT) 1. We have used in each case 90% od dataset for training and
10% for the testing and validation. After conducting our experiments, we decided to choose
the XML-Roberta-large model as the baseline model because it performed better than all other
monolingual models. The performance disparity observed between certain monolingual models,
such as IndicBERT or Bangla Hate BERT, when compared to XML-Roberta-large, can potentially
be attributed to the fact that these models are not consistently up-to-date. Besides, observing
that XML-Roberta-large consistently beats three monolingual models, we decided to use it as
the default model. Table 2 presents F1 scores for diferent models and methods by Language.</p>
        <p>Data Augmentation Process: In our study, we employed data augmentation technique to
enhance the robustness and diversity of our dataset. We utilized ChatGPT-3, a state-of-the-art
language model to perform data augmentation, ensuring that our text samples maintained their
original annotation labels, specifically distinguishing between ”HOF” (Hate Speech or Ofensive)
and ”Not HOF” (Non-Hate Speech or Non-Ofensive). Our approach aimed to generate additional
samples while preserving the integrity of the original labels. To conduct this operation, we
selected ChatGPT-3 as our data augmentation tool due to its impressive language generation
capabilities and its ability to produce coherent and contextually relevant text. We formulated
a specific prompt for ChatGPT-3 to guide its generation of augmented samples. The prompt
played a critical role in instructing the model to maintain the original annotation label while
generating new content. An example of our prompt is as follows:</p>
        <p>“‘ Given the following text sample: [Original Text Sample], please generate three additional
samples that preserve the original annotation label (HOF or NOT). “‘</p>
        <p>This prompt structure ensured that ChatGPT-3 produces three augmented versions of each
input text while respecting the initial annotation. To assess the efectiveness of our data
augmentation process, we manually evaluated a random sample of 200 augmented results
generated by ChatGPT-3. During this evaluation, we compared the original annotation labels
with those of the augmented samples. We found that a remarkable 98% of the augmented
1https://github.com/Meem007/Hate-Speech-and-Ofensive-Content-Indentification
samples accurately preserve the original label. Table 3 presents example sentences generated
by ChatGPT3.</p>
        <sec id="sec-4-2-1">
          <title>Expanding the Dataset Through Self-Model Annotation:</title>
          <p>To further bolster our model’s performance, we targeted exclusively Bengali language due to
the unavailability of suitable datasets for Assamese and Bodo. For this purpose, we implemented
a complementary strategy aimed at enlarging our training dataset by incorporating additional
data. We leveraged an additional Bengali public dataset containing 30,000 samples from a
diferent domain, which we annotated using our initial baseline model 2.</p>
          <p>The approach involved the initially training the model using 90% of the existing training
data and reserving 10% for testing &amp; evaluation purposes. Subsequently, during the evaluation
phase, we established optimal thresholds and applied both upper and lower thresholds to
automatically label a portion of the test data. For example, we utilized an upper threshold of 0.90
and a lower threshold of 0.20. Following the automatic annotation of this portion of the new
public data using these thresholds, we retrained the model by incorporating this annotated data
into the training dataset. Remarkably, this process led to a 2.1% improvement in the model’s
performance.</p>
          <p>Our initial model employed a threshold of 0.50 for classification decisions. Recognizing the
potential for improved results by varying these thresholds, we embarked on an empirical journey
to explore higher and lower threshold values. It is important to emphasize that our threshold
selection process was not grounded in established scientific methodologies but rather conducted
as an experiment guided by intuition. We acknowledged that the chosen threshold values lacked
rigorous scientific validation. Nonetheless, the results we obtained were empirically favorable.
2Source: https://www.kaggle.com/datasets/naurosromim/bengali-hate-speech-dataset</p>
          <p>With this annotation technique, we were able to label an additional 5,000 samples out of the
30,000 in the public Kaggle dataset. However, it is important to note that, due to shared task
guideline, we did not include these results as part of our final submission.
"বােলর িশক্ষা মন্ত্রী প্রশাসিনক কােজ িনেয়ািজত।"</p>
          <p>Translation</p>
          <p>Label</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>4. Conclusion</title>
      <p>In addition to demonstrating the eficiency of the transformer-based models in identifying
abusive language, our research has illuminated the significance of model choice, task-specific
modifications, and creative approaches needed to tackle the challenges posed by multilingual and
cross-lingual scenarios. Our primary focus was on Task 4, which involves detecting ofensive
language in Assamese, Bengali, and Bodo languages. To handle the inherent complexities of
these tasks, we leveraged powerful transformer-based models, including XML-Roberta, among
others.</p>
      <p>Through rigorous assessments, a consistent trend emerged: XML-Roberta-large is found
to consistently outperform monolingual models across the majority of test scenarios. This
observation underscores the eficacy of employing cross-lingual pre-trained models, particularly
in resource-constrained settings.</p>
      <p>Furthermore, our data augmentation strategy, enabled by ChatGPT-3, has proven to be
remarkably successful in enlarging our dataset’s size and diversity, while preserving the
reliability of the original annotation labels. Notably, we observed substantial improvements of
nearly 1.2-1.5% across almost all three languages after augmentation. As part of future work, it
would be intriguing to compare this approach with other large language models (LLMs) and
existing traditional methods of text augmentation.</p>
      <p>To enhance the performance of our model, we expanded our training dataset exclusively for
Bengali using Self-Model annotation. We annotated this data with the help of our baseline model
and conducted experiments with various classification thresholds. Despite the lack of scientific
validation for threshold selection, this approach resulted in a noteworthy 2.1% performance
improvement. This technique shows a promise for domain adaptation and bolstering model
robustness. It is intriguing that, even though the model already had knowledge of this additional
data, however, adding self-annotated data further enhanced its learning, highlighting the
feasibility and potential benefits of such experiments.</p>
      <p>Our top-ranking results in Task 4 for Bengali and Assamese underscore the versatility and
reliability of these models in addressing real-world challenges. In conclusion, our research may
make a substantial contribution to the formidable task of identifying objectionable language
in multilingual and cross-lingual contexts. We anticipate that our findings will inspire further
scholarly investigations in this field, fostering the development of more potent and reliable
methods for detecting and mitigating ofensive language in online discourse.</p>
    </sec>
    <sec id="sec-6">
      <title>5. Acknowledgments</title>
      <p>We want to convey our appreciation to the HASOC-2023 organizers for afording us the chance
to engage in this shared task and for their assistance throughout the duration of the event.
Furthermore, we recognize the valuable contributions of our entire research team and the
resources made available by the NLP community, which were instrumental in facilitating this
research.
problem of ofensive language, in: Proceedings of the international AAAI conference on
web and social media, volume 11, 2017, pp. 512–515.
[13] B. Gambäck, U. K. Sikdar, Using convolutional neural networks to classify hate-speech, in:</p>
      <p>Proceedings of the first workshop on abusive language online, 2017, pp. 85–90.
[14] A. Joulin, E. Grave, P. Bojanowski, T. Mikolov, Bag of tricks for eficient text classification,
arXiv preprint arXiv:1607.01759 (2016).
[15] Z. Waseem, Are you a racist or am i seeing things? annotator influence on hate speech
detection on twitter, in: Proceedings of the first workshop on NLP and computational
social science, 2016, pp. 138–142.
[16] I. A. Farha, W. Magdy, Multitask learning for arabic ofensive language and hate-speech
detection, in: Proceedings of the 4th workshop on open-source Arabic corpora and
processing tools, with a shared task on ofensive language detection, 2020, pp. 86–90.
[17] A. Baruah, K. Das, F. Barbhuiya, K. Dey, Aggression identification in english, hindi and
bangla text using bert, roberta and svm, in: Proceedings of the second workshop on
trolling, aggression and cyberbullying, 2020, pp. 76–82.
[18] M. R. Karim, B. R. Chakravarthi, J. P. McCrae, M. Cochez, Classification benchmarks for
under-resourced bengali language based on multichannel convolutional-lstm network, in:
2020 IEEE 7th International Conference on Data Science and Advanced Analytics (DSAA),
IEEE, 2020, pp. 390–399.
[19] M. S. Jahan, M. Haque, N. Arhab, M. Oussalah, Banglahatebert: Bert for abusive language
detection in bengali, in: Proceedings of the Second International Workshop on Resources
and Techniques for User Information in Abusive Language Analysis, 2022, pp. 8–15.
[20] N. Romim, M. Ahmed, H. Talukder, M. Saiful Islam, Hate speech detection in the bengali
language: A dataset and its baseline evaluation, in: Proceedings of International Joint
Conference on Advances in Computational Intelligence: IJCACI 2020, Springer, 2021, pp.
457–468.
[21] K. Ghosh, D. Sonowal, A. Basumatary, B. Gogoi, A. Senapati, Transformer-based hate
speech detection in assamese, in: 2023 IEEE Guwahati Subsection Conference (GCON),
IEEE, 2023, pp. 1–5.
[22] T. Ranasinghe, K. Ghosh, A. S. Pal, A. Senapati, A. E. Dmonte, M. Zampieri, S. Modha,
S. Satapara, Overview of the HASOC subtracks at FIRE 2023: Hate speech and ofensive
content identification in assamese, bengali, bodo, gujarati and sinhala, in: Proceedings of
the 15th Annual Meeting of the Forum for Information Retrieval Evaluation, FIRE 2023,
Goa, India. December 15-18, 2023, ACM, 2023.
[23] M. Zampieri, S. Malmasi, P. Nakov, S. Rosenthal, N. Farra, R. Kumar, Semeval-2019 task 6:
Identifying and categorizing ofensive language in social media (ofenseval), arXiv preprint
arXiv:1903.08983 (2019).
[24] M. Zampieri, P. Nakov, S. Rosenthal, P. Atanasova, G. Karadzhov, H. Mubarak,
L. Derczynski, Z. Pitenis, Ç. Çöltekin, Semeval-2020 task 12: Multilingual ofensive
language identification in social media (ofenseval 2020), arXiv preprint arXiv:2006.07235
(2020).
[25] T. Mandl, S. Modha, A. Kumar M, B. R. Chakravarthi, Overview of the hasoc track at fire
2020: Hate speech and ofensive language identification in tamil, malayalam, hindi, english
and german, in: Proceedings of the 12th Annual Meeting of the Forum for Information
Association, Marseille, France, 2022, pp. 7–12. URL: https://aclanthology.org/2022.wildre-1.
2.
[38] S. Sarker, Banglabert: Bengali mask language model for bengali language understanding,
2020. URL: https://github.com/sagorbrur/bangla-bert.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>M. S.</given-names>
            <surname>Jahan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Oussalah</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. K.</given-names>
            <surname>Mim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Islam</surname>
          </string-name>
          ,
          <article-title>Ofensive language identification using hindi-english code-mixed tweets, and code-mixed data augmentation</article-title>
          ,
          <source>in: Working Notes of FIRE 2021-Forum for Information Retrieval Evaluation</source>
          , Gandhinagar, India,
          <source>December 13-17</source>
          ,
          <year>2021</year>
          , RWTH Aachen University,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>J. T.</given-names>
            <surname>Nockleby</surname>
          </string-name>
          , Hate speech,
          <source>Encyclopedia of the American constitution 3</source>
          (
          <year>2000</year>
          )
          <fpage>1277</fpage>
          -
          <lpage>1279</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>P.</given-names>
            <surname>Burnap</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. L.</given-names>
            <surname>Williams</surname>
          </string-name>
          ,
          <article-title>Cyber hate speech on twitter: An application of machine classification and statistical modeling for policy and decision making</article-title>
          ,
          <source>Policy &amp; internet 7</source>
          (
          <year>2015</year>
          )
          <fpage>223</fpage>
          -
          <lpage>242</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>Y. J.</given-names>
            <surname>Foong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Oussalah</surname>
          </string-name>
          ,
          <article-title>Cyberbullying system detection and analysis</article-title>
          ,
          <source>in: 2017 European Intelligence and Security Informatics Conference (EISIC)</source>
          , IEEE,
          <year>2017</year>
          , pp.
          <fpage>40</fpage>
          -
          <lpage>46</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>R.</given-names>
            <surname>Kumar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. K.</given-names>
            <surname>Ojha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Malmasi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Zampieri</surname>
          </string-name>
          ,
          <article-title>Benchmarking aggression identification in social media</article-title>
          ,
          <source>in: Proceedings of the first workshop on trolling, aggression and cyberbullying (TRAC-2018)</source>
          ,
          <year>2018</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>11</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>K.</given-names>
            <surname>Reynolds</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Kontostathis</surname>
          </string-name>
          ,
          <string-name>
            <surname>L. Edwards,</surname>
          </string-name>
          <article-title>Using machine learning to detect cyberbullying</article-title>
          ,
          <source>in: 2011 10th International Conference on Machine learning and applications and workshops</source>
          , volume
          <volume>2</volume>
          , IEEE,
          <year>2011</year>
          , pp.
          <fpage>241</fpage>
          -
          <lpage>244</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>K.</given-names>
            <surname>Dinakar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Jones</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Havasi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Lieberman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Picard</surname>
          </string-name>
          ,
          <article-title>Common sense reasoning for detection, prevention, and mitigation of cyberbullying</article-title>
          ,
          <source>ACM Transactions on Interactive Intelligent Systems (TiiS) 2</source>
          (
          <issue>2012</issue>
          )
          <fpage>1</fpage>
          -
          <lpage>30</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Bounab</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Adeegbe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Oussalah</surname>
          </string-name>
          ,
          <article-title>Towards storytelling automatic textual summerized</article-title>
          ,
          <source>Proceedings of the FRUCT'25</source>
          , Helsinki, Finland,
          <fpage>5</fpage>
          -8
          <source>November</source>
          <year>2019</year>
          (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>K.</given-names>
            <surname>Ghosh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Senapati</surname>
          </string-name>
          , U. Garain,
          <article-title>Baseline bert models for conversational hate speech detection in code-mixed tweets utilizing data augmentation and ofensive language identification in marathi, in: Forum for Information Retrieval Evaluation (Working Notes)(FIRE)</article-title>
          .
          <source>CEUR-WS. org</source>
          ,
          <year>2022</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>K.</given-names>
            <surname>Ghosh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. A.</given-names>
            <surname>Senapati</surname>
          </string-name>
          ,
          <article-title>Hate speech detection: a comparison of mono and multilingual transformer model with cross-language evaluation</article-title>
          ,
          <source>in: Proceedings of the 36th Pacific Asia Conference on Language, Information and Computation</source>
          , De La Salle University, Manila, Philippines,
          <year>2022</year>
          , pp.
          <fpage>853</fpage>
          -
          <lpage>865</lpage>
          . URL: https://aclanthology.org/
          <year>2022</year>
          .paclic-
          <volume>1</volume>
          .
          <fpage>94</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>M. S.</given-names>
            <surname>Jahan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Oussalah</surname>
          </string-name>
          ,
          <article-title>A systematic review of hate speech automatic detection using natural language processing</article-title>
          .,
          <string-name>
            <surname>Neurocomputing</surname>
          </string-name>
          (
          <year>2023</year>
          )
          <fpage>126232</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>T.</given-names>
            <surname>Davidson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Warmsley</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Macy</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Weber</surname>
          </string-name>
          ,
          <source>Automated hate speech detection and the Retrieval Evaluation</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>29</fpage>
          -
          <lpage>32</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [26]
          <string-name>
            <given-names>S.</given-names>
            <surname>Modha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Mandl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G. K.</given-names>
            <surname>Shahi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Madhu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Satapara</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Ranasinghe</surname>
          </string-name>
          ,
          <string-name>
            <surname>M.</surname>
          </string-name>
          <article-title>Zampieri, Overview of the hasoc subtrack at fire 2021: Hate speech and ofensive content identification in english and indo-aryan languages and conversational hate speech</article-title>
          ,
          <source>in: Proceedings of the 13th Annual Meeting of the Forum for Information Retrieval Evaluation</source>
          ,
          <year>2021</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>3</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [27]
          <string-name>
            <given-names>R.</given-names>
            <surname>Raj</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Srivastava</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Saumya</surname>
          </string-name>
          , Nsit &amp; iiitdwd@ hasoc
          <year>2020</year>
          :
          <article-title>Deep learning model for hate-speech identification in indo-european languages</article-title>
          .,
          <source>in: FIRE (Working Notes)</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>161</fpage>
          -
          <lpage>167</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [28]
          <string-name>
            <given-names>R.</given-names>
            <surname>Kumar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Lahiri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. K.</given-names>
            <surname>Ojha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Bansal</surname>
          </string-name>
          , Comma@ fire
          <year>2020</year>
          :
          <article-title>Exploring multilingual joint training across diferent classification tasks</article-title>
          .,
          <source>in: FIRE (Working Notes)</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>823</fpage>
          -
          <lpage>828</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [29]
          <string-name>
            <given-names>S.</given-names>
            <surname>Satapara</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Madhu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Ranasinghe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. E.</given-names>
            <surname>Dmonte</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Zampieri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Pandya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Shah</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Sandip</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Majumder</surname>
          </string-name>
          , T. Mandl,
          <article-title>Overview of the hasoc subtrack at fire 2023:hate-speech identification in sinhala and gujarati</article-title>
          , in: K. Ghosh,
          <string-name>
            <given-names>T.</given-names>
            <surname>Mandl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Majumder</surname>
          </string-name>
          , M. Mitra (Eds.), Working Notes of FIRE 2023 -
          <article-title>Forum for Information Retrieval Evaluation, Goa, India</article-title>
          .
          <source>December 15-18</source>
          ,
          <year>2023</year>
          , CEUR Workshop Proceedings, CEUR-WS.org,
          <year>2023</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [30]
          <string-name>
            <given-names>H.</given-names>
            <surname>Madhu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Satapara</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Pandya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Shah</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Mandl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Modha</surname>
          </string-name>
          ,
          <article-title>Overview of the hasoc subtrack at fire 2023: Identification of conversational hate-speech, in:</article-title>
          K. Ghosh,
          <string-name>
            <given-names>T.</given-names>
            <surname>Mandl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Majumder</surname>
          </string-name>
          , M. Mitra (Eds.), Working Notes of FIRE 2023 -
          <article-title>Forum for Information Retrieval Evaluation, Goa, India</article-title>
          .
          <source>December 15-18</source>
          ,
          <year>2023</year>
          , CEUR Workshop Proceedings, CEUR-WS.org,
          <year>2023</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [31]
          <string-name>
            <given-names>S.</given-names>
            <surname>Masud</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. A.</given-names>
            <surname>Khan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. S.</given-names>
            <surname>Akhtar</surname>
          </string-name>
          , T. Chakraborty,
          <article-title>Overview of the HASOC Subtrack at FIRE 2023: Identification of Tokens Contributing to Explicit Hate in English by Span Detection</article-title>
          , in: Working Notes of FIRE 2023 -
          <article-title>Forum for Information Retrieval Evaluation</article-title>
          ,
          <string-name>
            <surname>CEUR</surname>
          </string-name>
          ,
          <year>2023</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [32]
          <string-name>
            <given-names>K.</given-names>
            <surname>Ghosh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Senapati</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. S.</given-names>
            <surname>Pal</surname>
          </string-name>
          ,
          <article-title>Annihilate Hates (Task 4</article-title>
          ,
          <string-name>
            <surname>HASOC</surname>
          </string-name>
          <year>2023</year>
          )
          <article-title>: Hate Speech Detection in Assamese, Bengali, and Bodo languages</article-title>
          , in: Working Notes of FIRE 2023 -
          <article-title>Forum for Information Retrieval Evaluation</article-title>
          ,
          <string-name>
            <surname>CEUR</surname>
          </string-name>
          ,
          <year>2023</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [33]
          <string-name>
            <given-names>S.</given-names>
            <surname>Satapara</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Masud</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Madhu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. A.</given-names>
            <surname>Khan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. S.</given-names>
            <surname>Akhtar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Chakraborty</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Modha</surname>
          </string-name>
          , T. Mandl,
          <article-title>Overview of the HASOC subtracks at FIRE 2023: Detection of hate spans and conversational hate-speech, in: Proceedings of the 15th Annual Meeting of the Forum for Information Retrieval Evaluation</article-title>
          ,
          <string-name>
            <surname>FIRE</surname>
          </string-name>
          <year>2023</year>
          , Goa,
          <source>India. December 15-18</source>
          ,
          <year>2023</year>
          , ACM,
          <year>2023</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [34]
          <string-name>
            <given-names>J.</given-names>
            <surname>Devlin</surname>
          </string-name>
          , M.-
          <string-name>
            <given-names>W.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Toutanova</surname>
          </string-name>
          , Bert:
          <article-title>Pre-training of deep bidirectional transformers for language understanding</article-title>
          , arXiv preprint arXiv:
          <year>1810</year>
          .
          <volume>04805</volume>
          (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [35]
          <string-name>
            <given-names>A.</given-names>
            <surname>Conneau</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Khandelwal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Goyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Chaudhary</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Wenzek</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Guzmán</surname>
          </string-name>
          , E. Grave,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ott</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Zettlemoyer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Stoyanov</surname>
          </string-name>
          ,
          <article-title>Unsupervised cross-lingual representation learning at scale</article-title>
          , arXiv preprint arXiv:
          <year>1911</year>
          .
          <volume>02116</volume>
          (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [36]
          <string-name>
            <given-names>D.</given-names>
            <surname>Kakwani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Kunchukuttan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Golla</surname>
          </string-name>
          ,
          <string-name>
            <surname>G. N.C.</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Bhattacharyya</surname>
          </string-name>
          ,
          <string-name>
            <surname>M. M. Khapra</surname>
          </string-name>
          , P. Kumar, IndicNLPSuite: Monolingual Corpora,
          <article-title>Evaluation Benchmarks and Pre-trained Multilingual Language Models for Indian Languages</article-title>
          , in: Findings of EMNLP,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [37]
          <string-name>
            <given-names>R.</given-names>
            <surname>Nayak</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Joshi</surname>
          </string-name>
          ,
          <article-title>L3Cube-HingCorpus and HingBERT: A code mixed Hindi-English dataset and BERT language models</article-title>
          ,
          <source>in: Proceedings of the WILDRE-6 Workshop within the 13th Language Resources and Evaluation Conference</source>
          , European Language Resources
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>