=Paper=
{{Paper
|id=Vol-3681/T6-14
|storemode=property
|title=Examining Hate Speech Detection Across Multiple Indo-Aryan Languages in Tasks 1 & 4
|pdfUrl=https://ceur-ws.org/Vol-3681/T6-14.pdf
|volume=Vol-3681
|authors=Gyandeep Kalita,Eisha Halder,Chetna Taparia,Advaitha Vetagiri,Dr. Partha Pakray
|dblpUrl=https://dblp.org/rec/conf/fire/KalitaHTVP23
}}
==Examining Hate Speech Detection Across Multiple Indo-Aryan Languages in Tasks 1 & 4==
Examining Hate Speech Detection Across Multiple Indo-Aryan Languages in Tasks 1 & 4 Gyandeep Kalita1,† , Eisha Halder1,† , Chetna Taparia1,† , Advaitha Vetagiri1,* and Dr. Partha Pakray1 1 National Institute of Technology Silchar Abstract Hate speech continues to be a pressing concern in online social media (OSM) platforms, necessitating effective automated detection systems. In this paper, we propose a unified approach, encompassing both Task 1 & 4, to tackle the challenge of hate speech recognition within the HASOC 2023 framework. It addresses the complexities of multilingual OSM by employing cutting-edge Natural Language Processing (NLP) techniques and leveraging powerful language models put forward by team CNLP-NITS-PP. The key objective is optimising precision-recall trade-offs in hate speech detection, spanning English and Indo-Aryan languages. The empirical results demonstrate the effectiveness of our approach in isolating explicit signs of hate speech, emphasizing model efficiency, interpretability, and the importance of diverse linguistic nuances in creating safer online environments. This integrated work sets the stage for advancements in hate-span detection and underlines the significance of fostering responsible and inclusive online conversations across various language environments. Keywords Online social media, Multilingual, Natural Language Processing, CNN, BiLSTM, BERT, GPT-2, Named Entity Recognition. 1. Introduction Social media platforms such as Twitter and Facebook have become integral to modern life, providing a global platform for individuals to express themselves. However, the openness of these platforms has also led to the proliferation of harmful content, including hate speech and harassment [1]. This has underscored the need for automated systems to identify and address abusive language in online conversations [2] [3]. Detecting offensive content is challenging due to its diverse linguistic forms, necessitating context-aware models to pinpoint hateful or abusive text snippets [4]. Additionally, implicit forms of hate speech require the deduction of pragmatic implications [5]. The spread of hate speech and inflammatory language on social media platforms is a major worldwide problem in today’s digital age, as communication plays a crucial role in determining Forum for Information Retrieval Evaluation, December 15-18, 2023, India * Corresponding author. † These authors contributed equally for Task 1 & 4. $ gyandeepkalita1@gmail.com (G. Kalita); eishashalder@gmail.com (E. Halder); chetna.taparia@gmail.com (C. Taparia); advaitha21\protect1_rs@cse.nits.ac.in (A. Vetagiri); partha@cse.nits.ac.in (Dr. P. Pakray) https://parthapakray.com/ (Dr. P. Pakray) 0000-0002-0651-4171 (A. Vetagiri); 0000-0003-3834-5154 (Dr. P. Pakray) © 2023 Copyright for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0). CEUR Workshop Proceedings http://ceur-ws.org ISSN 1613-0073 CEUR Workshop Proceedings (CEUR-WS.org) CEUR ceur-ws.org Workshop ISSN 1613-0073 Proceedings public debate. Low-resource languages like Sinhala, Gujarati, Bengali, Bodo, and Assamese, which have received little attention in the field of Natural Language Processing (NLP), are severely affected by this problem. Our research activities cover a range of tasks for identifying harmful and hateful content in these underrepresented languages1 . In Task 1, we tackle Sinhala (Task 1A), a language with a unique alphabet and intricate grammatical structures, and further broaden our emphasis to Gujarati (Task 1B), where a dearth of labelled data poses a significant obstacle. Task 4 extends the study’s horizons by including Bengali, Bodo, and Assamese [6] [7]. These languages, which are rich in cultural richness and legacy, have generally been disregarded in NLP research, especially when it comes to the identification of hate speech. Our study uses statistics painstak- ingly gathered from social media sites to use binary classification to characterize material as hate/offensive or not. The importance of our work lies in its role in safeguarding cultural identities and developing secure online environments for these language speakers. To respond to the complexities of hate speech in different linguistic and cultural contexts, we use cutting-edge NLP approaches, language-specific feature engineering, and pre-processing. Additionally, we investigate how models developed for languages with abundant resources may be applied to languages with limited resources to improve hate speech identification. Through this extensive study project, we hope to advance responsible digital communication, better understand how to identify hate speech in different linguistic contexts, and create a more welcoming online space for every language. 2. Literature Review Hate speech detection in fairly low-resourced languages such as Sinhala and regional Indian languages has recently attracted research attention [8]. With the proliferation of user-generated content on social media platforms, there is an urgent need to identify and moderate hateful and offensive content in these regional languages (Mathew et al., 2021)[9]. For the Sinhala Language, a few research works have been conducted. (Munasinghe et al., 2022)[10] contributed an annotated dataset of Sinhala Tweets annotated into Hate or Non-Hate. They also developed and compared the performance of different architectures such as CNN, LSTM, BiGRU and an ensemble of various other Deep Learning architectures. (Sandaruwan et al., 2019)[11] also contributed a labeled dataset containing texts from Facebook and YouTube for Hate Detection in Sinhala and compared classification results using simple Machine Learning Classifiers such as SVMs, MNB, RFDT, etc. The SOLD: Sinhala Offensive Language Dataset, a labeled dataset for Offensive content detection in Sinhala, was contributed by (Ranasinghe et al.,2022)[12], which also forms the basis for the dataset provided for the Task 1A of HASOC 2023. In case of Indian Languages, prior work on hate speech detection has concentrated primarily on Hindi and Malayalam. For instance, (Bohra et al., 2018)[13] presented a dataset for hate speech identification in Hindi-English code-mixed social media text. They tested various classification models, including fastText, CNN, GRU, and LSTM 1 Github Repository For Gujarati, (Khurana et al., 2022)[14] contributed a novel model to detect hate comments in 13 Indian languages that included Gujarati based on XLM-RoBERTa (XLM-R) using the Moj Multilingual Abusive Comment Identification dataset. For the Bengali language, (Das et al., 2020)[15] compiled a labelled dataset of YouTube comments for Bengali hate speech recognition. They compared machine learning models like SVM, NB and deep learning architectures like CNN, GRU, and capsule networks. Assamese is a relatively low-resource language. (Ghosh et al., 2023)[16] contributed a dataset for binary hate classification in Assamese and described an approach for hate detection using various BERT models. For the Bodo language, there is limited prior research. The HASOC 2023 shared task provides the pioneering benchmark hate speech detection dataset in Bodo. This will encourage further research in this low-resource language (Chakravarthi et al., 2021)[17] (Vetagiri et al., 2023a)[18] leveraged GPT-2 to automatically classify online sexist content. Their work demonstrates the potential of sizable pre-trained language models for hate speech detection. In another work, (Vetagiri et al., 2023b)[4] proposed an approach using CNN-BiLSTM and domain-specific embeddings for online sexism prediction. Much previous work has relied on machine learning and deep neural networks. But these necessitate substantial labelled datasets, which are scarce for low-resource languages. Recent emphasis has focused on multilingual models such as mBERT, which can leverage data from high-resource languages. Domain adaptation approaches have also proven effective in adapting models trained on English data. The HASOC 2023 shared tasks furnishes standard labeled benchmark datasets for hate speech detection. This will catalyze research in these languages and progress the state-of-the-art. Multilingual models and cross-lingual transfer learning are promising avenues to explore for these languages. 3. Dataset and Task Description 3.1. Tasks Description: Task 1 of the HASOC’23 aimed at identifying hate, offensive, and profane content in social media posts in two languages, namely Sinhala(Task 1a) and Gujarati(Task 1b) [19]. Task 4 was similar to Task 1 and required us to detect hate speech in three other Indian languages, Bengali, Assamese and Bodo. For all the given languages, the training and test datasets had already been provided. Creating coarse-grained binary classification models to divide tweets into the following two categories was the primary goal for the tasks: • Hate and Offensive(HOF): Posts that contain hate speech, vulgarity, or offensive material. • Non-Hate and Offensive (NOT): Posts devoid of offensive language, hate speech, or any other negative elements. 3.2. Data Source 3.2.1. Sinhala Dataset (Task 1a) The Sinhala dataset provided for train and test had been sourced from the recently released SOLD: Sinhala Offensive Language Detection dataset, which served as a comprehensive resource for the particular task. The training dataset had been further divided into three columns. The first one consisted of the post id, the second of the tweet text, while the third column consisted of the labels, HOF and NOT, for each of the corresponding tweets in the same row. 3.2.2. Gujarati Dataset(Task 1b) For Task 1b, the training dataset had 200 tweets, primarily categorized into two labels, HOF and NOT, besides three other columns, including the tweet id, the UserName and the date of creation. It is noteworthy that the exact source of the dataset has not been mentioned in the materials provided for the competition. 3.2.3. Assamese, Bengali and Bodo Dataset(Task 4) The training and test datasets for the task had already been provided in all three languages, Assamese Bengali and Bodo. However, it is worth noting that none of the sources for the data were explicitly mentioned. 3.3. Data Statistics Distribution of HASOC’23 training datasets for Task 1 and Task 4. For each language, the total no of text entries and the corresponding no of tweets per class are shown below. Table 1 Task 1 & 4 dataset stastistics. Language Total text entries Hate and Offensive(HOT) Not Hate(NOT) Assamese 4035 2346 1689 Bengali 2180 515 1665 Bodo 1678 998 680 Sinhala 7500 3176 4324 Gujarati 200 100 100 Besides this, any external use of data beyond what was provided required explicit permission from the competition organizers. Test Set size: For Task 1A, the Sinhala test dataset consisted of 2500 tweets. This had to be labeled as either HOF or NOT based on our model. For Task 1B, the Gujarati test dataset consisted of a total of 1196 tweets to be labelled similarly. For Task 4, the Bengali, Assamese, and Bodo test datasets consisted of 320, 1009, and 420 text entries, respectively. These entries had to be labeled as either HOF or NOT based on our model. 3.4. Data Preprocessing We used a number of standard preprocessing methods prior to training our model using the given datasets. Given that the training datasets provided had their texts sourced from Twitter, it was anticipated to contain certain unwanted elements, such as emojis, URLs, mentions and special characters. In order to guarantee the accuracy and relevancy of the text data, we followed procedures to remove such unwanted noise. 4. Methodology In this section, we describe the methodology and the experimental setup used for the various tasks under HASOC 2023. We conducted a thorough investigation into various neural network architectures, pretrained Large Language models, and classical machine learning models to identify the most effective model for the task. 4.1. Task 1: Identifying Hate, offensive and profane content in Sinhala & Gujarati For the task of Identifying Hate, offensive and profane content in Sinhala & Gujarati, the models that resulted in the best performance are as follows: • A CNN-based Binary Classification Model with FastText Embeddings. • A CNN-BiLSTM based Hybrid Model with FastText/GloVE Embeddings 4.1.1. CNN + FastText Binary Classification Model : Inspired by the works of Kim et al., [20], we developed the model based on the CNN architecture. At the core of our model lies the input layer, where text sequences representing individual posts are processed. To prepare the input data, we concatenate the words within each sentence, with the sequence length capped at 70 words. The words here are represented as dense vectors of 300 dimensions using pre-trained FastText embeddings for the respective languages. Using machine learning or related dimensional reduction techniques, word embedding converts each token into a vector of real numbers to quantify and classify the semantic similarity of linguistic phrases based on their distributional qualities in a large corpus. For the convolutional layer, we employed a one-dimensional convolution operation utilizing 100 filters with a kernel size of 3, leading to a systematic scanning of the text sequences and identifying pertinent patterns in the data. An activation function, the rectified linear unit (ReLU), was also applied to introduce non-linearity and enable the model to capture complex relationships in the data. Subsequently, a dense layer with 50 neurons and a ReLU activation function, coupled with an L2-norm constraint, was added to transform the extracted features further. Dropout with a rate of 0.5 was applied as a regularization technique to prevent overfitting. The resultant vector was then concatenated with the feature vector and the output was passed onto a dense output layer with sigmoid activation and cross entropy loss as shown in Figure 1, to produce the binary hate classification for the model. 4.1.2. CNN-BiLSTM + FastText/GLoVE Binary Classification Model : Based on the contributions of Vetagiri et al., [4], we developed the model, which is a combination of two different model architectures - the Convolutional Neural Networks (CNN) (Kim, 2014)[20] layer for identifying local textual patterns in the input text and Bidirectional Long Short-Term Memory (BiLSTM) Liu and Guo et al., [21] layer as a form of the Recurrent Neural Architecture Sherstinsky et al., [22] for understanding the long-term complex sequential dependencies within the text data. The output of these two layers is then passed through a dense layer with a sigmoid activation function for Binary Classification. To prepare the Input data, the model uses the exact pre- trained FastText embeddings for the respective language mentioned above, representing the words as 300-dimensional dense vectors, with the sequence length capped at 70 words, which was held as non-trainable. A similar implementation of this model using pre-trained GLoVE embeddings Kumar et al., [23] showed identical results. For the CNN layers, we first employed a SpatialDropout1D layer, a dropout variant that selectively drops entire 1D feature maps during training, to combat overfitting. Subsequently, a one-dimensional convolution layer with 64 filters and a kernel size of 3 was used to capture local textual patterns with fine granularity. For the BiLSTM part, we used the initial layer with 128 units and a return sequence setting with a dropout of 0.1 and recurrent dropout of 0.1 to process the text inputs in both forward as well as reverse directions followed by a Global Average Pooling and a dense layer with 128 units and a rectified linear unit (ReLU) activation function to introduce non-linearity. Subsequently, a dropout layer is employed whose output is then concatenated with the feature vector and passed through a dense layer with sigmoid activation as shown in figure 2 to produce the overall model for Binary Hate Classification. Our models are trained using the RMSprop optimiser, and our loss function is a binary cross-entropy function. To fine-tune our hyper-parameters over a range of values, we conduct a grid search and select the best-performing model according to validity accuracy. However, no attempt at experimentation by reversing the order of the CNN & Bi-LSTM Layers was made for this particular task. 4.2. Task 4: Identifying Hate, offensive and profane content in Bengali, Bodo, and Assamese languages The two model architectures used in Task 1 were also experimented with in Task 4. These architectures were used to implement the CNN and the CNN-BiLSTM models which used pre-trained FastText embeddings in the respective languages (except for Bodo, for which Hindi embeddings were used) representing each word as a dense vector with 300 dimensions. In addition to these two architectures, several others were also experimented with, the details for which are discussed below: Figure 1: CNN + FastText Binary Classification Figure 2: CNN-BiLSTM + FastText/GLoVE Binary Model Architecture Classification Model Architecture 4.2.1. Pre-trained BERT Architecture Considering the low-resource nature of Task 4 and the limited size of the datasets, we exper- imented with pre-trained models based on the Bidirectional Encoder Representations from Transformers (BERT) architecture Devlin et al., [24] [25]. We experimented with the Tensorflow Hub to access the pre-trained BERT models. We used the "bert-multi-cased-preprocess/3" for text processing and the "bert-multi-cased-L- 12-H-768-A-12/4" encoder for contextualized word embeddings from the TensorFlow Hub, which is trained on multilingual Wikipedia Data. The model utilizes a BERT preprocessing layer for tokenization and embeddings of the input text, followed by a BERT encoder layer to generate contextual embeddings containing the complex contextual relationship within the language. This is followed by a dropout Neural Layer with a 10% dropout rate to enhance generalization and mitigate overfitting. The final trainable dense layer with 769 employs the sigmoid activation function, producing the binary classification outputs for the given languages 4.2.2. GPT-2 Model We also explored GPT-2 as a state-of-the-art pre-trained large language model for the task. GPT-2 is a transformer-based model that takes a sequence of words, represented as dense vectors, as input and uses many intermediate layers to extract contextual information for the input text. The output is then passed through a dense layer, producing the Binary Classifier. For the task, we used a pre-trained GPT-2 that contained 768 parameters, fine-tuned on the training dataset for each language in an 80-20% split and a further 20% split from the training set for validation. The input text is tokenised and passed through the model for fine-tuning. Figure 3: Pre-trained BERT Figure 4: GPT-2 Model Architecture Model Architecture The model uses an Adam optimizer for optimization with a learning rate of 1e-5. The batch size for the model is 8 4.2.3. Classical Machine-Learning Based Models Due to the small size of the training datasets provided for all three languages, we tried implement- ing classical Machine-Learning approaches as well to accomplish our goals. We experimented with multiple Machine-Learning architectures such as Support Vector Machines (SVMs), Linear Regression, Logistic Regression and Random Forest Classifiers. However, although the training set accuracies and the macro F1 scores for all the above architectures were quite close, we observed that the Logistic Regression Model achieved the best overall performance. Hence, we created a simple Logistic Regression Model using SciKitLearn Library, trained the Model with the training datasets for each language, and implemented simple tf-idf vectorisation for embedding the input sentences 5. Results and Analysis 5.1. Task 1 & 4 For evaluation the models, the test accuracy for this test data was used for the initial evaluation of the models. In addition to this, the Macro F1 scores, which were acquired by the models on submission in the HASOC 2023 Submission portal were also considered. For Task 4, although the performance of each of the aforementioned models was analyzed, only a few of the models for each language gave the best performance, summarized in Tables 1 & 2. As it is evident from Tables 1 & 2, the CNN-BiLSTM+FastText/GLoVe Model gave the best Macro F1 Score for Task 1A(Sinhala), and the CNN+FastText Model gave the best Macro F1 Table 2 Performance scores for Tasks 1A & 1B in terms of their test accuracies and Macro F1 scores Task Language Model Accuracy Macro F1 1A Sinhala CNN+FastText 0.7800 0.7556 CNN-BiLSTM+FastText 0.7781 0.7711 1B Gujarati CNN+FastText 0.7025 0.6873 CNN-BiLSTM+FastText 0.7121 0.6758 Table 3 Performance scores for Task 4 in terms of their test accuracies Language Model Accuracy Macro F1 CNN+FastText 0.6381 0.60108 CNN-BiLSTM+FastText/GLoVe 0.6122 Bengali CNN+FastText+ External Dataset 0.8177 GPT-2 0.5924 BERT 0.5992 Logistic Regression 0.6325 CNN+FastText 0.6374 0.59485 Assamese CNN-BiLSTM+FastText/GLoVe 0.6183 GPT-2 0.5825 Logistic Regression 0.6559 CNN+FastText 0.6577 Bodo CNN-BiLSTM+FastText/GLoVe 0.6162 Logistic Regression 0.6755 0.66925 Score for Task 1B(Gujarati). For Task 4, in both Bengali and Assamese, the CNN+FastText Model gave the highest accuracy with a Macro F1 score of 0.60108 in Bengali and 0.59485 in Assamese. For Bodo, evidently, the simple Logistic Regression Model gave a much better performance than the other model architectures with a Macro F1 score of 0.66925, possibly due to the small size of the training dataset. 6. Conclusion and Future Scope Our research has been dedicated to the vital task of identifying hate speech, particularly in Indo-Aryan languages such as Bengali, Assamese, Bodo, Gujarati, and Sinhala. We’ve devised a comprehensive strategy that unites these linguistic intricacies under a single versatile model. Through domain-aware pre-training and meticulous alignment of our models with language-specific context, we’ve significantly enhanced hate speech detection. Furthermore, our exploration into model ensemble techniques has bolstered detection accuracy and resilience across diverse language settings, laying a foundational step towards comprehensive hate speech detection in Indo-Aryan languages. Our overarching goal is to foster a safer and more inclusive digital space for speakers of diverse linguistic backgrounds. As we look to the future, our work paves the way for further advancements in hate-span detection, focusing on model efficiency, interpretability, and an expansive training data corpus encompassing evolving hate speech trends and linguistic variations. We also recognize the potential of real-time monitoring and context-aware integration in dynamically evolving online environments. Acknowledgments We wish to extend our appreciation to the Computer Science and Engineering Department of the National Institute of Technology Silchar for granting us the opportunity to carry out our research and experiments. We are grateful for the support, resources, and research environment offered by the CNLP & AI Lab at NIT Silchar. References [1] B. Vidgen, L. Derczynski, (2020), Directions in abusive language training data, a systematic review: Garbage in, garbage out. PloS one 15 (2020). [2] P. Fortuna, S. Nunes, A survey on automatic detection of hate speech in text, ACM Computing Surveys (CSUR) 51 (2018) 1–30. [3] S. Satapara, S. Masud, H. Madhu, M. A. Khan, M. S. Akhtar, T. Chakraborty, S. Modha, T. Mandl, Overview of the HASOC subtracks at FIRE 2023: Detection of hate spans and conversational hate-speech, in: Proceedings of the 15th Annual Meeting of the Forum for Information Retrieval Evaluation, FIRE 2023, Goa, India. December 15-18, 2023, ACM, 2023. [4] A. Vetagiri, P. Adhikary, P. Pakray, A. Das, CNLP-NITS at SemEval-2023 task 10: Online sexism prediction, PREDHATE!, in: Proceedings of the 17th International Workshop on Semantic Evaluation (SemEval-2023), Association for Computational Linguistics, Toronto, Canada, 2023, pp. 815–822. URL: https://aclanthology.org/2023.semeval-1.113. doi:10. 18653/v1/2023.semeval-1.113. [5] D. Jurgens, L. Hemphill, E. Chandrasekharan, A just and comprehensive strategy for using NLP to address online abuse, Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, 2019. [6] K. Ghosh, A. Senapati, A. S. Pal, Annihilate Hates (Task 4, HASOC 2023): Hate Speech Detection in Assamese, Bengali, and Bodo languages, in: Working Notes of FIRE 2023 - Forum for Information Retrieval Evaluation, CEUR, 2023. [7] T. Ranasinghe, K. Ghosh, A. S. Pal, A. Senapati, A. E. Dmonte, M. Zampieri, S. Modha, S. Satapara, Overview of the HASOC subtracks at FIRE 2023: Hate speech and offensive content identification in assamese, bengali, bodo, gujarati and sinhala, in: Proceedings of the 15th Annual Meeting of the Forum for Information Retrieval Evaluation, FIRE 2023, Goa, India. December 15-18, 2023, ACM, 2023. [8] K. Ghosh, D. A. Senapati, Hate speech detection: a comparison of mono and multilingual transformer model with cross-language evaluation, in: Proceedings of the 36th Pacific Asia Conference on Language, Information and Computation, De La Salle University, Manila, Philippines, 2022, pp. 853–865. URL: https://aclanthology.org/2022.paclic-1.94. [9] B. Mathew, P. Saha, S. M. Yimam, C. Biemann, P. Goyal, A. Mukherjee, Hatexplain: A benchmark dataset for explainable hate speech detection, in: Proceedings of the AAAI conference on artificial intelligence, volume 35, 2021, pp. 14867–14875. [10] S. Munasinghe, U. Thayasivam, A deep learning ensemble hate speech detection approach for sinhala tweets, in: 2022 Moratuwa Engineering Research Conference (MERCon), 2022, pp. 1–6. doi:10.1109/MERCon55799.2022.9906232. [11] H. Sandaruwan, S. Lorensuhewa, M. Kalyani, Sinhala hate speech detection in social media using text mining and machine learning, in: 2019 19th International Conference on Advances in ICT for Emerging Regions (ICTer), volume 250, 2019, pp. 1–8. doi:10.1109/ ICTer48817.2019.9023655. [12] T. Ranasinghe, I. Anuradha, D. Premasiri, K. Silva, H. Hettiarachchi, L. Uyangodage, M. Zampieri, Sold: Sinhala offensive language dataset, 2022. arXiv:2212.00851. [13] A. Bohra, D. Vijay, V. Singh, S. S. Akhtar, M. Shrivastava, A dataset of hindi-english code-mixed social media text for hate speech detection, in: Proceedings of the second workshop on computational modeling of people’s opinions, personality, and emotions in social media, 2018, pp. 36–41. [14] R. Khurana, C. Pandey, P. Gupta, P. Nagrath, AniMOJity:detecting hate comments in Indic languages and analysing bias against content creators, in: Proceedings of the 19th International Conference on Natural Language Processing (ICON), Association for Computational Linguistics, New Delhi, India, 2022, pp. 172–182. URL: https://aclanthology. org/2022.icon-main.23. [15] A. Das, N. Tandon, S. Narayan, HASoC: Hate Speech and Offensive Content Identification in Indo-European Languages: Overview of HASoC Track at FIRE 2020, in: Forum for Information Retrieval Evaluation (FIRE), 2020, pp. 29–32. [16] K. Ghosh, D. Sonowal, A. Basumatary, B. Gogoi, A. Senapati, Transformer-based hate speech detection in assamese, in: 2023 IEEE Guwahati Subsection Conference (GCON), 2023, pp. 1–5. doi:10.1109/GCON58516.2023.10183497. [17] B. R. Chakravarthi, R. Priyadharshini, N. Jose, A. Kumar M, T. Mandl, P. K. Kumaresan, R. Ponnusamy, H. R L, J. P. McCrae, E. Sherly, Findings of the shared task on offensive language identification in Tamil, Malayalam, and Kannada, in: Proceedings of the First Workshop on Speech and Language Technologies for Dravidian Languages, Association for Computational Linguistics, Kyiv, 2021, pp. 133–145. URL: https://aclanthology.org/2021. dravidianlangtech-1.17. [18] A. Vetagiri, P. K. Adhikary, P. Pakray, A. Das, “Leveraging GPT-2 for Automated Classifi- cation of Online Sexist Content“, In Exist 2023 Lab at CLEF 2023: Conference and Labs of the Evaluation Forum, September 18–21, 2023, Thessaloniki, Greece, 2023. [19] S. Satapara, H. Madhu, T. Ranasinghe, A. E. Dmonte, M. Zampieri, P. Pandya, N. Shah, M. Sandip, P. Majumder, T. Mandl, Overview of the hasoc subtrack at fire 2023: Hate- speech identification in sinhala and gujarati, in: K. Ghosh, T. Mandl, P. Majumder, M. Mitra (Eds.), Working Notes of FIRE 2023 - Forum for Information Retrieval Evaluation, Goa, India. December 15-18, 2023, CEUR Workshop Proceedings, CEUR-WS.org, 2023. [20] Y. Kim, Convolutional neural networks for sentence classification, 2014. arXiv:1408.5882. [21] G. Liu, J. Guo, Bidirectional lstm with attention mechanism and convolutional layer for text classification, Neurocomputing 337 (2019) 325–338. URL: https:// www.sciencedirect.com/science/article/pii/S0925231219301067. doi:https://doi.org/ 10.1016/j.neucom.2019.01.078. [22] A. Sherstinsky, Fundamentals of recurrent neural network (rnn) and long short-term memory (lstm) network, Physica D: Nonlinear Phenomena 404 (2020) 132306. URL: https://www.sciencedirect.com/science/article/pii/S0167278919305974. doi:https://doi. org/10.1016/j.physd.2019.132306. [23] S. Kumar, S. Kumar, D. Kanojia, P. Bhattacharyya, “a passage to India”: Pre-trained word embeddings for Indian languages, in: Proceedings of the 1st Joint Workshop on Spoken Language Technologies for Under-resourced languages (SLTU) and Collaboration and Computing for Under-Resourced Languages (CCURL), European Language Resources association, Marseille, France, 2020, pp. 352–357. URL: https://aclanthology.org/2020.sltu-1. 49. [24] J. Devlin, M. Chang, K. Lee, K. Toutanova, BERT: pre-training of deep bidirectional transformers for language understanding, CoRR abs/1810.04805 (2018). URL: http://arxiv. org/abs/1810.04805. arXiv:1810.04805. [25] K. Ghosh, A. Senapati, U. Garain, Baseline bert models for conversational hate speech detection in code-mixed tweets utilizing data augmentation and offensive language identifi- cation in marathi, in: Fire, 2022. URL: https://api.semanticscholar.org/CorpusID:259123570.