<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Ofensive Language Identification on Multilingual code-mixed Text using BERT</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Snehaan Bhawal</string-name>
          <email>mailtosnehaan@gmail.com</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Pradeep Kumar Roy</string-name>
          <email>pradeep.roy@iiitsurat.ac.in</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Abhinav Kumar</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Indian Institute of Information Technology Surat</institution>
          ,
          <addr-line>Gujarat</addr-line>
          ,
          <country country="IN">India</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Kalinga Institute of Industrial Technology</institution>
          ,
          <addr-line>Odisha</addr-line>
          ,
          <country country="IN">India</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Siksha 'O' Anusandhan, Deemed to be University</institution>
          ,
          <addr-line>Bhubaneswar, Odisha</addr-line>
          ,
          <country country="IN">India</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2021</year>
      </pub-date>
      <abstract>
        <p>Hate Speech and Ofensive Content detection in social media has been an active field of research for the last couple of years. For the majority of the world consisting of non-native English speakers, most of the time unoficial messages are written in code-mixed language in a combination of words in a native language with English text. The current study focuses on using Machine and Deep learning techniques for detection of Hate Speech and Ofensive content in a Malayalam and Tamil code-mixed text collected from social media. The study showed that Deep learning models perform better than the machine learning models, specifically the implementation of BERT based transfer learning models performed best.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Hate Speech is generally defined as content that expresses hate or prejudice against a particular
group, ethnicity, religion, nationality or sexual orientation [
        <xref ref-type="bibr" rid="ref1 ref2 ref3">1, 2, 3</xref>
        ]. Social network platforms
consists of a large amount of user-generated content, and due to being not moderated in nature,
there is a widespread use of targeted hate speech against certain individuals, which has become
a very critical issue [
        <xref ref-type="bibr" rid="ref4 ref5">4, 5</xref>
        ].
      </p>
      <p>
        Humans can’t always moderate the social media networks to read, identify, and deal with the
hateful text that the platform generates in such high frequency afecting the users mentally.
Thus, there is a need for automation, and it has already been established that detection of
such content by automation has been successful to a certain extent. Davidson et al. [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] used
Logistic Regression with n-grams TF-IDF features to perform classification of Ofensive and
Non-Ofensive text. At the same time, in another paper, a neural network-based approach was
presented by Badjatiya et al. [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], where they used GloVe embedding with CNNs and LSTMs to
provide better results.
      </p>
      <p>
        However, most of the research that has taken place over hate speech and ofensive language
detection is predominately for the English language [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. In a country like India, with home
nEvelop-O
(A. Kumar)
to numerous regional languages, people have adapted to using a mix of regional and English
languages to express themselves in social media [
        <xref ref-type="bibr" rid="ref8 ref9">8, 9</xref>
        ]. The current research is done upon
bilingual texts, which contain words from both languages and are written in one script, called
code-mixed text. While there is another way of combining the words in native writing with the
English script, which is known as Script-mixed text. These are far more challenging to work
with as it requires a diferent tokenization process compared to what we need for English texts.
Examples of some popular code-mixed languages in India are Hinglish (Hindi and English),
Tanglish (Tamil and English), Manglish (Malayalam and English), and a mixed language of
Kannada and English [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ].
      </p>
      <p>
        Identifying Hate Speech in such code-mixed languages is much more challenging than in
English [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] due to the absence of suficient NLP resources. The models which are trained on a
monolingual corpus might find it dificult to provide satisfactory results. This is because the
system learns and recognizes the words provided in the given vocabulary while training. In
the case of code-mixed text, many new words will be introduced which will not be present in
the training vocabulary. The words are then marked as out of vocabulary token that makes no
diference in the estimation of the model. Thus, the performance of the model decreases.
      </p>
      <p>
        The current study focuses on Ofensive language identification in code-mixed languages of
Tanglish and Manglish with the data set provided in HASOC-Dravidian-CodeMix-FIRE2021
challenge. An overview of the dataset can be found here [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. We have implemented a number
of Machine learning and Deep learning models, including transfer learning models like BERT,
to distinguish between the Ofensive and Non-Ofensive text.
      </p>
      <p>The rest of the article is summarized as follows: Section 2 discusses the related works. Section
3, 3.1, 4 provides the task description, the pre-processing steps taken, followed by the explanation
of the proposed methodology. The experimental results and discussion are explained in Section
5 and 6, respectively. Section 7 concluded the work by highlighting the limitations and future
scope.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Literature Review</title>
      <p>
        The use of Hate speech and Ofensive language has become one of the major issues concerning
the social networking platforms and hence received fruitful attention from many worldwide
researchers [
        <xref ref-type="bibr" rid="ref1 ref2 ref4 ref8 ref9">1, 2, 4, 8, 9</xref>
        ]. Roy et al. [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] developed a deep learning-based framework to address
the hate speech issue on Twitter. They used a Convolutional Neural Network to process the
tweets and predict whether it were Hate or non-Hate. They considered only the tweets written
in English language and hence unable to detect the tweets of multilingual texts, such as
TamilEnglish, Kannada-English and others. Badjatiya et al. [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] developed a deep learning model to
classify the tweets into racist, sexist or neither category. Their model experimented on 16k
labelled data and outperformed existing models. The main issues with the existing works are
the coverage of the language. Most of the existing researches use an English dataset. However,
currently, people prefer to post the message on the social platform in code-mixed languages
like Hindi-English mixed, Tamil-English mixed and others.
      </p>
      <p>
        Recent work by Kumar et al. [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] suggested a deep learning-based framework to classify the
Tamil and Malayalam code-mixed YouTube comments into the ofensive and non-ofensive
categories. Many machines and deep learning models have experimented. The best result was
obtained when a character n-gram tf-idf features passed to the dense neural network. Their
model achieved the weighted F1-score value of 0.95. Suryawanshi et al. [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] developed the
resources for Tamilmeme detection. The developed dataset consisted of two labels: troll and
not_troll. A total of ten models were submitted, and the model with an F1-score value of 0.55
secured the first rank among them. Banerjee et al. [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] compared the performance of the
pre-trained models on the Hinglish code-mixed dataset for predicting the Hate and non-Hate
post.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Task and Data Description</title>
      <p>The current study is an implementation and comparison of diferent Machine and Deep Learning
models for a Hate Speech and Ofensive Language detection system for Tamil and Malayalam
code-mixed texts in English. The dataset consists of sentences collected from comments or posts
from social media. Table 1 shows the overview of the data used in this analysis. There are two
sets of data, Malayalam code-mixed and Tamil code-mixed data each consisting of code-mixed
sentences with addition of various emojis in most of the cases.</p>
      <sec id="sec-3-1">
        <title>3.1. Data Preprocessing</title>
        <p>As the data was code-mixed with Malayalam or Tamil mixed with English, no stop-word
removal was done. The text being informal in nature contained emojis and emoticons which
were replaced with their respective textual meaning using data from the Unicode Consortium’s
emoji code repository by using the demoji library. This was then followed with the removal of
punctuation, URLS, email-ids, hyperlinks and numeric data from the text.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Methodology</title>
      <p>This section discusses the working of the implemented models in detail, the codes of which can
be found in the GitHub repository1. In our current study, three diferent approaches were used
as shown in Figure 1:
i Conventional Machine learning based models.
ii Neural Network based models
iii Transfer learning based models</p>
      <sec id="sec-4-1">
        <title>4.1. Traditional ML Models</title>
        <p>
          In traditional ML-based models, we looked into using a 1 - 5 gram word TF-IDF feature set. The
extracted features were then fed to classifiers like Logistic Regression (LR), Naive Bayes (NB),
Random Forest (RF), XGBoost (XGB), and Support Vector Machine (SVM). The performance
of these models were evaluated in terms of precision, recall, and F1-score [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ]. The detailed
performance report of these models are provided in the Section 5.
        </p>
        <p>1https://github.com/Sbhawal/HASOC-FIRE-2021-CODES</p>
        <sec id="sec-4-1-1">
          <title>Conventional ML</title>
        </sec>
        <sec id="sec-4-1-2">
          <title>Models: RF NB XGB</title>
          <p>LR
SVM
rk Dense (128)
o
w
t
e
lN Dense (256)
a
r
u
N Dense (512)
e</p>
          <p>Dense (64)
N
N Dense (128)
C</p>
        </sec>
        <sec id="sec-4-1-3">
          <title>Conv1D</title>
        </sec>
        <sec id="sec-4-1-4">
          <title>Embedding Layer</title>
        </sec>
        <sec id="sec-4-1-5">
          <title>Preprocessing</title>
        </sec>
        <sec id="sec-4-1-6">
          <title>Code-Mixed</title>
        </sec>
        <sec id="sec-4-1-7">
          <title>Social Data</title>
        </sec>
        <sec id="sec-4-1-8">
          <title>Transfer Learning</title>
        </sec>
        <sec id="sec-4-1-9">
          <title>Models:</title>
        </sec>
        <sec id="sec-4-1-10">
          <title>BERT</title>
        </sec>
        <sec id="sec-4-1-11">
          <title>Indic BERT</title>
        </sec>
        <sec id="sec-4-1-12">
          <title>MuRIL</title>
        </sec>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Neural Network based models</title>
        <p>In neural network-based models, the 1 to 5 grams TF-IDF features extracted while working with
the Machine Learning models were used again as an input to a simple deep neural network
(DNN) model. This model consisted of four fully connected layers in sequential order, with 512,
256, 128 and 1 neurons in the first, second, third and fourth (output) layers. Due to classification
between two distinct labels, only one output neuron was used to identify the outputs. The
hidden neurons were set up with the ReLU activation function. In contrast, the output neuron was
set up with sigmoid activation function with Adam and binary-cross-entropy as the respective
chosen optimizer and loss function.</p>
        <p>
          The second experimented neural model is Convolutional Neural Network (CNN) [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ]. The
CNN consisted of one Conv1D layer followed by a Global Max Pooling and a Dropout layer
connected to a fully connected sequential network with two hidden layers of 128 and 64 neurons,
respectively, The activation function for the hidden neurons were chosen to be ReLU. The output
was a single neuron with sigmoid activation. An embedding layer was used as the input layer
with the embedding dimension set to 50 and the input length set to 120. Therefore, a (120,
50) dimensional embedding matrix was given as an input to CNN. The Convolutional layer
consisted of 64 filters with a kernel size of three.
        </p>
        <p>Our final neural network based model was a Bidirectional Long Short-Term Memory model
(Bi-LSTM), consisting of 256 memory units followed by Global Max Pooling and Batch
Normalization. The input layer was an embedding layer with 50 dimensions and length padded
to 120 like the previous model. Two fully connected dense layers served as the hidden layers
comprising of 20 and 10 neurons, respectively with ReLU as the activation function, which was
then connected to a single neuron as the output layer with sigmoid activation.
Subsequent Hyper-parameter tuning was done for the described models to check for the
optimal performance by adjusting the optimizer, learning rate and embedding dimensions. Our
experiments led to the best result with a learning rate set to 0.0001 with the optimizer set
as Adam. The embedding dimension was set to 50 as it gave the best result. Due to binary
classification and the overall balanced nature of the data set, the loss function was kept to be
binary cross-entropy and sigmoid activation function for the output neuron.</p>
      </sec>
      <sec id="sec-4-3">
        <title>4.3. Transfer Learning</title>
        <p>This study has implemented BERT (Bidirectional Encoder Representations from Transformers)
models to work with these models’ transfer learning capabilities. For these models, no
preprocessing was done. Three diferent variants of BERT models were studied.</p>
        <p>
          i BERT (multilingual) [
          <xref ref-type="bibr" rid="ref17">17</xref>
          ].
ii IndicBERT [
          <xref ref-type="bibr" rid="ref18">18</xref>
          ].
iii Multilingual Representations for Indian Languages (MuRIL) [
          <xref ref-type="bibr" rid="ref19">19</xref>
          ].
        </p>
        <p>
          The BERT [
          <xref ref-type="bibr" rid="ref17">17</xref>
          ] multilingual model was trained on 102 languages with masked language
modelling. The case-sensitive model was chosen, as no prior data pre-processing was done
in case of transfer learning models. IndicBERT [
          <xref ref-type="bibr" rid="ref18">18</xref>
          ] is a multilingual ALBERT model,
pretrained exclusively on a corpus of 12 major Indian languages. Compared to other such BERT
based models, IndicBERT is comparatively smaller and has much less number of parameters.
We used ktrain [
          <xref ref-type="bibr" rid="ref20">20</xref>
          ] libraries to develop the IndicBERT model. The last model that we used is
MuRIL(Multilingual Representations for Indian Languages) [
          <xref ref-type="bibr" rid="ref19">19</xref>
          ]. MuRIL is a BERT model trained
over a monolingual corpus of 17 Indian languages along with their translated and transliterated
counterpart. The diferentiating factor between this and the previous model is that IndicBERT
is trained only on the native Indian scripts. In contrast, MuRIL is trained on traditional scripts
as well their transliterated corpus in roman script. The benefit of this will be evident in our
experiment, which deals with code-mixed data of Indian and English language written strictly
in roman script.
        </p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Results</title>
      <p>Ofensive
Not Ofensive
Weighted Avg
Ofensive
Not Ofensive
Weighted Avg
Ofensive
Not Ofensive
Weighted Avg
Ofensive
Not Ofensive
Weighted Avg
Ofensive
Not Ofensive
Weighted Avg
This section presents the results of all our experiments done during this study, as mentioned in
Section 2. The results shown below corresponds to the model prediction on the validation data
and are shown in terms of precision, recall, and F1-score belonging to OFF (Ofensive) or NOT
(Not Ofensive) class. A model is said to be the best if it reports the highest weighted average in
terms of precision, recall, and F1-score. The best results for the particular data set are presented
in bold for each diferent model used in this study.</p>
      <p>Traditional ML models were built using 1 to 5-gram character TF-IDF features which included
the following models, LR, RF, NB, XGB and SVM. Their results are shown in Table 2 respectively.
In the Malayalam code-mixed data set, the LR classifier gave a better performance with recall
and F1 of 0.70. Similarly, in Tamil code-mixed text, the LR classifier performed the best and
reported precision of 0.83 with recall and an F1-score of 0.82.</p>
      <p>Results of the neural network models are presented in Table 3. It is seen that a simple DNN
provided the best results in the case of the Malayalam Code Mix data with a precision of 0.75
with recall and F1-Score being 0.74. In Tamil Code Mix data, CNN showed the best performance
with precision reaching 0.90 with Recall and F1-Score of 0.89.</p>
      <p>In Table 4 the results of using diferent BERT models are presented. In both Malayalam and
Tamil Data, it is seen that the MuRIL model performed the best among the other models. In
Malayalam data, the precision was 0.79 with recall and F1-Score being 0.78, and for Tamil data,
precision, recall and F1-Score were 0.91, which was the highest among all experimented models</p>
    </sec>
    <sec id="sec-6">
      <title>6. Discussion</title>
      <p>
        Among all experimented models, the MuRIL-a transfer learning model, performed the best
for both Malayalam and Tamil code-mixed data. The experimental outcomes show that the
traditional machine learning models are unable to understand the context of the message and
hence may not be a good choice for this task. A simple Deep Neural Network (DNN) with
an embedding layer performed better than most of the machine learning models (Tables 2,
3). Although some of their performance came near those of neural network models, we were
dealing mostly with text data consisting of single sentences. For multiple sentence texts, a
neural network with the ability to hold some memory like LSTM would have outclassed the
machine learning models [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ].
      </p>
      <p>As shown in Table 4, the IndicBERT model is not able to perform as good as the multilingual
BERT model. This may happen because the data set consisted of code-mixed data in the Roman
script only. If there were any text written in the traditional script, then the multilingual BERT
model would have treated most of the tokens as an unknown token which would had afected
the model performance—benefiting the IndicBERT model as it was trained on monolingual
Indian scripts. Finally, MuRIL, which was trained on a corpus of both traditional script and
transliterated one, performed better than all the models.</p>
      <p>The above reported results (Tables 2, 3, 4) were based on the predictions done over the
validation data set. While using the test data, the proposed MuRIL model achieved the precision,
recall and F1-score value of 0.679, 0.673 and 0.636, respectively for Tamil code-mixed data, while
on the Malayalam code-mixed data, the precision, recall and F1-score value is 0.752, 0.727, and
0.734, respectively for the best case.</p>
      <p>The models were re-experimented with labelled test data, and the obtained results with
diferent machine learning, neural network and transfer learning models are shown in Table 5.
Similar to the results on the validation data, MuRIL -a transfer learning model produces the
best prediction outcomes in terms of weighted average precision, recall and F1-score for both
Tanglish and Manglish test dataset.</p>
    </sec>
    <sec id="sec-7">
      <title>7. Conclusion</title>
      <p>Hate speech and ofensive language detection is still a challenge for low resource and
codemixed languages in NLP. We implemented various machine learning, deep learning and transfer
learning models to find the best suitable model for code-mixed Tamil and Malayalam datasets.
The results reported by the models show the deep learning models. Specifically, the pre-trained
models outperformed the machine learning models. The MuRIL model performed the best
reporting weighted F1-score of 0.636 in Tamil code-mixed data. The same model provided a
weighted F1-score of 0.734 in Malayalam code-mixed data. On test data, the BERT and MuRIL
both transfer learning model yielded almost similar outcomes. In the future, a better model
can be built by some additional preprocessing steps on the dataset to achieve better prediction
accuracy.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>P. K.</given-names>
            <surname>Roy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. K.</given-names>
            <surname>Tripathy</surname>
          </string-name>
          ,
          <string-name>
            <surname>T. K. Das</surname>
            ,
            <given-names>X.-Z.</given-names>
          </string-name>
          <string-name>
            <surname>Gao</surname>
          </string-name>
          ,
          <article-title>A framework for hate speech detection using deep convolutional neural network</article-title>
          ,
          <source>IEEE Access 8</source>
          (
          <year>2020</year>
          )
          <fpage>204951</fpage>
          -
          <lpage>204962</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>S.</given-names>
            <surname>Saumya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Kumar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. P.</given-names>
            <surname>Singh</surname>
          </string-name>
          ,
          <article-title>Ofensive language identification in dravidian code mixed social media text</article-title>
          ,
          <source>in: Proceedings of the First Workshop on Speech and Language Technologies for Dravidian Languages</source>
          ,
          <year>2021</year>
          , pp.
          <fpage>36</fpage>
          -
          <lpage>45</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>T.</given-names>
            <surname>Mandl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Modha</surname>
          </string-name>
          ,
          <string-name>
            <surname>A. Kumar</surname>
            <given-names>M</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>B. R.</given-names>
            <surname>Chakravarthi</surname>
          </string-name>
          ,
          <article-title>Overview of the hasoc track at ifre 2020: Hate speech and ofensive language identification in tamil, malayalam, hindi, english and german</article-title>
          ,
          <source>in: Forum for Information Retrieval Evaluation</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>29</fpage>
          -
          <lpage>32</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>A.</given-names>
            <surname>Kumar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Saumya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. P.</given-names>
            <surname>Singh</surname>
          </string-name>
          ,
          <article-title>Nitp-ai-nlp@ hasoc-dravidian-codemix-fire2020: A machine learning approach to identify ofensive languages from dravidian code-mixed text</article-title>
          .,
          <source>in: FIRE (Working Notes)</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>384</fpage>
          -
          <lpage>390</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>A.</given-names>
            <surname>Kumar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Saumya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. P.</given-names>
            <surname>Singh</surname>
          </string-name>
          ,
          <article-title>Nitp-ai-nlp@ hasoc-fire2020: Fine tuned bert for the hate speech and ofensive content identification from social media</article-title>
          .,
          <source>in: FIRE (Working Notes)</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>266</fpage>
          -
          <lpage>273</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>T.</given-names>
            <surname>Davidson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Warmsley</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Macy</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Weber</surname>
          </string-name>
          ,
          <article-title>Automated hate speech detection and the problem of ofensive language</article-title>
          ,
          <source>in: Proceedings of the International AAAI Conference on Web and Social Media</source>
          , volume
          <volume>11</volume>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>P.</given-names>
            <surname>Badjatiya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Gupta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Gupta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Varma</surname>
          </string-name>
          ,
          <article-title>Deep learning for hate speech detection in tweets</article-title>
          ,
          <source>in: Proceedings of the 26th international conference on World Wide Web companion</source>
          ,
          <year>2017</year>
          , pp.
          <fpage>759</fpage>
          -
          <lpage>760</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>S. M.</given-names>
            <surname>Jayanthi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Gupta</surname>
          </string-name>
          , Sj_aj@
          <fpage>dravidianlangtech</fpage>
          -eacl2021:
          <article-title>Task-adaptive pretraining of multilingual bert models for ofensive language identification</article-title>
          ,
          <source>arXiv preprint arXiv:2102.01051</source>
          (
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>C.</given-names>
            <surname>Vasantharajan</surname>
          </string-name>
          , U. Thayasivam, Hypers@
          <fpage>dravidianlangtech</fpage>
          -eacl2021:
          <article-title>Ofensive language identification in dravidian code-mixed youtube comments and posts</article-title>
          ,
          <source>in: Proceedings of the First Workshop on Speech and Language Technologies for Dravidian Languages</source>
          ,
          <year>2021</year>
          , pp.
          <fpage>195</fpage>
          -
          <lpage>202</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>B. R.</given-names>
            <surname>Chakravarthi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Priyadharshini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Jose</surname>
          </string-name>
          , A.
          <string-name>
            <surname>Kumar</surname>
            <given-names>M</given-names>
          </string-name>
          , T. Mandl,
          <string-name>
            <given-names>P. K.</given-names>
            <surname>Kumaresan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Ponnusamy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H. R L</given-names>
            ,
            <surname>J. P. McCrae</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Sherly</surname>
          </string-name>
          ,
          <article-title>Findings of the shared task on ofensive language identification in Tamil, Malayalam, and Kannada</article-title>
          ,
          <source>in: Proceedings of the First Workshop on Speech and Language Technologies for Dravidian Languages, Association for Computational Linguistics</source>
          , Kyiv,
          <year>2021</year>
          , pp.
          <fpage>133</fpage>
          -
          <lpage>145</lpage>
          . URL: https://aclanthology.org/
          <year>2021</year>
          . dravidianlangtech-
          <volume>1</volume>
          .
          <fpage>17</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>B. R.</given-names>
            <surname>Chakravarthi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. K.</given-names>
            <surname>Kumaresan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Sakuntharaj</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. K.</given-names>
            <surname>Madasamy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Thavareesan</surname>
          </string-name>
          , P. B,
          <string-name>
            <given-names>S. Chinnaudayar</given-names>
            <surname>Navaneethakrishnan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. P.</given-names>
            <surname>McCrae</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Mandl</surname>
          </string-name>
          ,
          <article-title>Overview of the HASOC-DravidianCodeMix Shared Task on Ofensive Language Detection in Tamil and Malayalam</article-title>
          , in: Working Notes of FIRE 2021 -
          <article-title>Forum for Information Retrieval Evaluation</article-title>
          ,
          <string-name>
            <surname>CEUR</surname>
          </string-name>
          ,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>R.</given-names>
            <surname>Priyadharshini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B. R.</given-names>
            <surname>Chakravarthi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Thavareesan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Chinnappa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Durairaj</surname>
          </string-name>
          , E. Sherly,
          <article-title>Overview of the dravidiancodemix 2021 shared task on sentiment detection in tamil, malayalam, and kannada, in: Forum for Information Retrieval Evaluation</article-title>
          ,
          <string-name>
            <surname>FIRE</surname>
          </string-name>
          <year>2021</year>
          ,
          <article-title>Association for Computing Machinery</article-title>
          ,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>S.</given-names>
            <surname>Suryawanshi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B. R.</given-names>
            <surname>Chakravarthi</surname>
          </string-name>
          ,
          <article-title>Findings of the shared task on troll meme classification in Tamil</article-title>
          ,
          <source>in: Proceedings of the First Workshop on Speech and Language Technologies for Dravidian Languages, Association for Computational Linguistics</source>
          , Kyiv,
          <year>2021</year>
          , pp.
          <fpage>126</fpage>
          -
          <lpage>132</lpage>
          . URL: https://aclanthology.org/
          <year>2021</year>
          .dravidianlangtech-
          <volume>1</volume>
          .
          <fpage>16</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>S.</given-names>
            <surname>Banerjee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B. Raja</given-names>
            <surname>Chakravarthi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. P.</given-names>
            <surname>McCrae</surname>
          </string-name>
          ,
          <article-title>Comparison of pretrained embeddings to identify hate speech in indian code-mixed text</article-title>
          ,
          <source>in: 2020 2nd International Conference on Advances in Computing, Communication Control and Networking (ICACCCN)</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>21</fpage>
          -
          <lpage>25</lpage>
          .
          <source>doi:1 0 . 1 1</source>
          <volume>0</volume>
          <fpage>9</fpage>
          <string-name>
            <surname>/ I C A C C C N</surname>
          </string-name>
          <article-title>5 1</article-title>
          <volume>0 5 2 . 2 0 2 0 . 9 3 6 2 7 3 1 .</volume>
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>D.</given-names>
            <surname>Tripathi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. R.</given-names>
            <surname>Edla</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Cheruku</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Kuppili</surname>
          </string-name>
          ,
          <article-title>A novel hybrid credit scoring model based on ensemble feature selection and multilayer ensemble classification</article-title>
          ,
          <source>Computational Intelligence</source>
          <volume>35</volume>
          (
          <year>2019</year>
          )
          <fpage>371</fpage>
          -
          <lpage>394</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>P. K.</given-names>
            <surname>Roy</surname>
          </string-name>
          ,
          <article-title>Multilayer convolutional neural network to filter low quality content from quora</article-title>
          ,
          <source>Neural Processing Letters</source>
          <volume>52</volume>
          (
          <year>2020</year>
          )
          <fpage>805</fpage>
          -
          <lpage>821</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>J.</given-names>
            <surname>Devlin</surname>
          </string-name>
          , M.-
          <string-name>
            <given-names>W.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Toutanova</surname>
          </string-name>
          , Bert:
          <article-title>Pre-training of deep bidirectional transformers for language understanding</article-title>
          , arXiv preprint arXiv:
          <year>1810</year>
          .
          <volume>04805</volume>
          (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>D.</given-names>
            <surname>Kakwani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Kunchukuttan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Golla</surname>
          </string-name>
          ,
          <string-name>
            <surname>G. N.C.</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Bhattacharyya</surname>
          </string-name>
          ,
          <string-name>
            <surname>M. M. Khapra</surname>
          </string-name>
          , P. Kumar, IndicNLPSuite: Monolingual Corpora,
          <article-title>Evaluation Benchmarks and Pre-trained Multilingual Language Models for Indian Languages</article-title>
          , in: Findings of EMNLP,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>S.</given-names>
            <surname>Khanuja</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Bansal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Mehtani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Khosla</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Dey</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Gopalan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. K.</given-names>
            <surname>Margam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Aggarwal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. T.</given-names>
            <surname>Nagipogu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Dave</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Gupta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. C. B.</given-names>
            <surname>Gali</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Subramanian</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Talukdar</surname>
          </string-name>
          , Muril:
          <article-title>Multilingual representations for indian languages</article-title>
          ,
          <year>2021</year>
          .
          <article-title>a r X i v : 2 1 0 3 . 1 0 7 3 0</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>A. S.</given-names>
            <surname>Maiya</surname>
          </string-name>
          ,
          <article-title>ktrain: A low-code library for augmented machine learning</article-title>
          ,
          <source>arXiv preprint arXiv:2004</source>
          .
          <volume>10703</volume>
          (
          <year>2020</year>
          ).
          <article-title>a r X i v : 2 0 0 4 . 1 0 7 0 3</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>P. K.</given-names>
            <surname>Roy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. P.</given-names>
            <surname>Singh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Banerjee</surname>
          </string-name>
          ,
          <article-title>Deep learning to filter sms spam</article-title>
          ,
          <source>Future Generation Computer Systems</source>
          <volume>102</volume>
          (
          <year>2020</year>
          )
          <fpage>524</fpage>
          -
          <lpage>533</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>