<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>Forum for Information Retrieval Evaluation, December</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Detection using Modified Cross-entropy Loss</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Arka Mitra</string-name>
          <email>thearkamitra@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Priyanshu Sankhala</string-name>
          <email>priyanshu.nitrr.ele@gmail.com</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Indian Institute of Technology</institution>
          ,
          <addr-line>Kharagpur</addr-line>
          ,
          <country country="IN">India</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>National Institute of Technology Raipur</institution>
          ,
          <country country="IN">India</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2021</year>
      </pub-date>
      <volume>1</volume>
      <fpage>3</fpage>
      <lpage>17</lpage>
      <abstract>
        <p>The number of increased social media users has led to a lot of people misusing these platforms to spread ofensive content and use hate speech. Manual tracking the vast amount of posts is impractical so it is necessary to devise automated methods to identify them quickly. Large language models are trained on a lot of data and they also make use of contextual embeddings. We fine-tune the large language models to help in our task. The data is also quite unbalanced; so we used a modified cross-entropy loss to tackle the issue. We observed that using a model which is fine-tuned in hindi corpora performs better. Our team (HNLP) achieved the macro F1-scores of 0.808, 0.639 in English Subtask A and English Subtask B respectively. For Hindi Subtask A, Hindi Subtask B our team achieved macro F1-scores of 0.737, 0.443 CEUR Workshop Proceedings</p>
      </abstract>
      <kwd-group>
        <kwd>Hate speech detection</kwd>
        <kwd>Text classification</kwd>
        <kwd>Deep-learning</kwd>
        <kwd>Transfer learning</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        With the increased use of social media platform like Twitter, Facebook, Instagram, and YouTube
by users around the world, the platforms have had positive aspects including but not limited
to social interaction, meeting like-minded people, giving a voice to each individual to share
their opinions [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. However, as a result, social media platforms can also be used to spread hate
comments, hate posts by certain individuals or groups; which can lead to having anxiety, mental
illness and severe stress to people who consume that hate content [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. It becomes necessary
to be able to detect such activities at its earliest to stop it from spreading, thereby making
social media a healthy place to interact and share their views without a fear of getting hate
comments[
        <xref ref-type="bibr" rid="ref3">3</xref>
        ].
      </p>
      <p>
        The hate posts can be insults or racist or discriminating on the bases of a particular gender,
religion, nationality, age bracket, ethnicity. Such comments can also lead to goading of violence
amongst people. With the large number of posts being shared each minute, it is not possible to
manually classify each of the posts. Thus, a pre-programmed system is required to distinguish
Hate speech activities quickly as hate content gains a lot of attention and is subject to be shared
fast as well [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. Direct targeted abuses and profane content are not that dificult to classify.
https://thearkamitra.github.io/ (A. Mitra); https://priyanshusankhala.github.io/ (P. Sankhala)
      </p>
      <p>
        © 2021 Copyright for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0).
However, it is extremely hard to recognize indirect hate content often involving use of humour,
irony, sarcasm even for an human annotator when the context of the posts are not provided.
This makes the classification task additionally more dificult for most progressive frameworks.
HASOC 2021 [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] is a shared task for the identification of hateful and ofensive content in English
and Indo-Aryan Languages. We participated in two sub-tasks for English and Hindi language
[
        <xref ref-type="bibr" rid="ref6">6</xref>
        ].
      </p>
      <p>The sub task A refers to classifying twitter samples into:
• H O F Hate and ofensive :- contains hate speech/profane/ofensive content.
• N O T Non Hate-ofensive :- which does not contain any hate speech, profane, ofensive
content.</p>
      <sec id="sec-1-1">
        <title>The sub task B refers to classifying twitter samples into:</title>
        <p>• H A T E Hate speech :- Posts under this class contain Hate speech content.
• O F F N Ofensive :- Posts under this class contain ofensive content.
• P R F N Profane :- These posts contain profane words.</p>
        <p>• N O N E Non-Hate :- These posts do not contain any hate speech content.</p>
        <p>For tasks pertaining to English language, we experimented with large language models like
ifne-tuning BERT (Bidirectional Encoder Representation from Transformer) [ 7], RoBERTa (A
Robustly Optimized BERT Pretraining Approach) [8] and XLNet (Generalized Autoregressive
Pretraining for Language Understanding) [9] out of which RoBERTa outperformed others with
the macro F1-score of 0.8089 while BERT and XLNet had the macro F1-score of 0.8050 and
0.7757 respectively in Subtask A and for Subtask B the macro F1-score was 0.6396 with RoBERTa
model respectively. For the tasks referring to Hindi language, the authors used a model which
is fine-tuned on detecting Hinglish sentiments [ 10] and had the macro F1-score of 0.7379 for
Subtask A and macro F1-score of 0.4431 for Subtask B.</p>
      </sec>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>In this section, we will discuss the previous state of the art methods proposed for detection of
hate speech. The use of BERT and other transfer learning algorithms, and deep neural models
based on LSTMs and CNNs tend to perform similar but better than traditional classifiers such
as SVM [11]. The number of papers, trying to automate Hate speech detection, that have been
published in Web of Science has been increasing exponentially [12]. Waseem et al. [13] have
classified hate speech into diferent categories and led to the Ofensive Language Identification
Dataset (OLID) [14].</p>
      <p>There has been work in diferent sub fields of abuse like in sexism [ 15, 16], cyberbullying [17],
trolling [18] and so on. There are hate comments in most of the social media sites like Youtube
[19], Instagram [20] which shows the importance of having a generalized Hate detection model
[13]. Work done by Yin et al. [21] gives an overall idea of the generalizability of the diferent
models that are present for hate speech detection. For the diferent models, the features from
the input that are used have a great impact on the performance. Xu et al. [22] showed that
part-of-speech tags are quite successful for improving the model; it is further improved by
considering the sentiment values [23]. The sentences in the online platforms do not always
follow the normal textual formats or correct spellings. Thus, Mehdad et al. [24] used a character
level encoding rather than using the word level encoding proposed by Meyer et al. [25]. The
type of architecture used also impacts on the performance on the model. Swamy et al. [26]
performed a comprehensive study that shows how diferent models perform and generalize.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Methodology</title>
      <p>of hate speech content.</p>
      <sec id="sec-3-1">
        <title>3.1. Languages</title>
        <p>
          HASOC 2021 [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] has been going on for two years now and a lot of diferent ways are uncovered to
detect hate content [27, 28]. This paper covers the use of large language models for classification
The Hate speech and Ofensive Content Identification in English and Indo-Aryan Languages
HASOC 2021 [
          <xref ref-type="bibr" rid="ref5 ref6">5, 6</xref>
          ] purposes two diferent tasks, in 3 diferent languages English, Hindi, Marathi.
        </p>
        <sec id="sec-3-1-1">
          <title>The authors participated in both tasks for English and Hindi languages.</title>
        </sec>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Task description</title>
        <p>The first task in all languages know as ”Subtask A” refers to a classification problem of twitter
samples which were labelled as H O F - Hate and ofensive content and N O T - Not hate and ofensive
content. The second task, know as ”Subtask B” refers to a classification of twitter samples which
were labelled as P R F N - Profane Words, H A T E - Hate speech, O F F N - Ofensive Content, and N O N E
Non-hate content. The detailed description of all columns present in a dataset is given in Table
1 and the number of twitter samples corresponding to each label is given in Table 2.</p>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. Approach</title>
        <p>The dataset that is provided in all the subtasks has an unequal number of samples per class.
The authors used large-language models since the models are trained on a large amount of data
and thus can understand the semantic structure of sentences and the tokens that are sent as
inputs to these models have a contextual embedding associated with them. The output of the
model is taken and then pooled. The resulting output is then passed through a linear layer and
a argmax is used to find the expected class of the sentence as shown in Figure. 1.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Results</title>
      <p>The authors submitted four groups of results Table 3 gives the final results for our submission.
The results has been evaluated on a test dataset, which is about one-third of the training data
size, using the Macro F1 scores.</p>
      <p>The experiments showed that large cased BERT performed the best followed by RoBERTa and
the lowest scores were obtained from the BERT base model. The maximum sequence length that
is used has a direct impact on the performance; with a larger length having a better performance,
with the training time increases at the same time.</p>
      <p>The methodology followed for both English and Hindi are the same, but the performance
obtained for the English subtask is quite better than that for the Hindi subtask. This shows that
the language models are pretty good in understanding the semantics for English but fail to do
so for a low resource language like Hindi. The modified cross-entropy loss provided a better F1
score as compared to training with equal importance given to all of the separate classes.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Experimental Details</title>
      <p>For English language we experimented with RoBERTa base pre-trained model [8], fine tuned
BERT large cased architecture[7], and XLNet [9]- all for the same configuration, i.e, max length
is set to 120, batch size to 8 and trained with 4 number of epochs. AdamW optimizer [29] with
an initial learning rate of 2e-5 is used for training. Similarly for hindi language tasks we used a
pre-trained model [10] from the Hugging face [30] library. The Max length has been set to 200,
batch size was 8, and number of epochs was set to 4.</p>
      <p>There is a trade-of between the accuracy and the total number of tokens. The amount of time
the model takes for training is proportional to the square of the number of tokens. As the
number of tokens increases, the amount of time increases. However, when we truncate the
maximum length, some of the information present in the sentence gets lost and the prediction
for the sentence might be wrong. We had to consider a trade-pof between the accuracy and
the time it takes for the model to train. For deciding the maximum sentence length, about 99%
percentile of number of tokens in sentences is considered. For generating predictions we made a
split of 90 % for training and 10 % validation to compare the performance of diferent models, for
each specific task and based on F1 scores of a particular epoch we updated the model weights.
The weights corresponding to the best validation scores have been selected for inferring the
test values. We observed that usually 3, 4 trained epochs had a higher F1 score.
For reproducibility, the codes have been uploaded to github 1. The random seed has been set to
42.</p>
    </sec>
    <sec id="sec-6">
      <title>6. Conclusions</title>
      <p>In this paper, we explain the shared tasks presented by HASOC in English and Indo-Aryan
languages. We used large language models which are pre-trained on large corpora for hate
speech detection tasks and to evaluate predictions by diferent models a validation dataset was
created. In future work, we hope to try out more diferent fine tuned models.</p>
    </sec>
    <sec id="sec-7">
      <title>Acknowledgments</title>
      <p>
        The authors would like to thank the organizers of Hate Speech and Ofensive Content
Identification in Indo-Aryan Languages 2021 [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] for conducting this data challenge. The authors
gratefully acknowledge google colab for providing GPU’s to do the computation. All pre-trained
models is based upon work supported by Hugging Face [30].
1https://github.com/priyanshusankhala/hasoc-hnlp
guages, in: Working Notes of FIRE 2021 - Forum for Information Retrieval Evaluation,
CEUR, 2021. URL: http://ceur-ws.org/.
[7] J. Devlin, M.-W. Chang, K. Lee, K. Toutanova, BERT: Pre-training of Deep Bidirectional
      </p>
      <p>Transformers for Language Understanding, in: NAACL, 2019.
[8] Y. Liu, M. Ott, N. Goyal, J. Du, M. Joshi, D. Chen, O. Levy, M. Lewis, L. Zettlemoyer,
V. Stoyanov, RoBERTa: A Robustly Optimized BERT Pretraining Approach, ArXiv
abs/1907.11692 (2019).
[9] Z. Yang, Z. Dai, Y. Yang, J. Carbonell, R. Salakhutdinov, Q. V. Le, XLNet: Generalized</p>
      <p>Autoregressive Pretraining for Language Understanding, in: NeurIPS, 2019.
[10] M. Bhange, N. Kasliwal, Hinglishnlp: Fine-tuned language models for hinglish sentiment
detection (2020).
[11] S. Modha, T. Mandl, P. Majumder, D. Patel, Tracking Hate in Social Media: Evaluation,</p>
      <p>Challenges and Approaches, SN Comput. Sci. 1 (2020) 105.
[12] M. A. Paz, J. Montero-Díaz, A. Moreno-Delgado, Hate speech: A systematized review,</p>
      <p>SAGE Open 10 (2020).
[13] Z. Waseem, T. Davidson, D. Warmsley, I. Weber, Understanding abuse: A typology of
abusive language detection subtasks, in: Proceedings of the First Workshop on Abusive
Language Online, Association for Computational Linguistics, Vancouver, BC, Canada,
2017, pp. 78–84. URL: https://aclanthology.org/W17-3012. doi:1 0 . 1 8 6 5 3 / v 1 / W 1 7 - 3 0 1 2 .
[14] M. Zampieri, S. Malmasi, P. Nakov, S. Rosenthal, N. Farra, R. Kumar, Predicting the type
and target of ofensive posts in social media, in: Proceedings of the 2019 Conference of
the North American Chapter of the Association for Computational Linguistics: Human
Language Technologies, Volume 1 (Long and Short Papers), Association for Computational
Linguistics, Minneapolis, Minnesota, 2019, pp. 1415–1420. URL: https://aclanthology.org/
N19-1144. doi:1 0 . 1 8 6 5 3 / v 1 / N 1 9 - 1 1 4 4 .
[15] Z. Waseem, D. Hovy, Hateful symbols or hateful people? predictive features for hate
speech detection on Twitter, in: Proceedings of the NAACL Student Research Workshop,
Association for Computational Linguistics, San Diego, California, 2016, pp. 88–93. URL:
https://aclanthology.org/N16-2013. doi:1 0 . 1 8 6 5 3 / v 1 / N 1 6 - 2 0 1 3 .
[16] S. Jaki, T. de Smedt, M. Gwózdz, R. Panchal, A. Rossa, G. D. Pauw, Online hatred of women
in the Incels.me forum, Journal of Language Aggression and Conflict (2019).
[17] M. Dadvar, D. Trieschnigg, R. Ordelman, F. de Jong, Improving cyberbullying detection
with user context, in: P. Serdyukov, P. Braslavski, S. O. Kuznetsov, J. Kamps, S. Rüger,
E. Agichtein, I. Segalovich, E. Yilmaz (Eds.), Advances in Information Retrieval, Springer
Berlin Heidelberg, Berlin, Heidelberg, 2013, pp. 693–696.
[18] R. Kumar, A. K. Ojha, M. Zampieri, S. Malmasi (Eds.), Proceedings of the First Workshop
on Trolling, Aggression and Cyberbullying (TRAC-2018), Association for Computational
Linguistics, Santa Fe, New Mexico, USA, 2018. URL: https://aclanthology.org/W18-4400.
[19] K. Dinakar, R. Reichart, H. Lieberman, Modeling the Detection of Textual Cyberbullying,
in: The Social Mobile Web, 2011.
[20] H. Zhong, H. Li, A. C. Squicciarini, S. M. Rajtmajer, C. Grifin, D. J. Miller, C. Caragea,
Content-Driven Detection of Cyberbullying on the Instagram Social Network, in: IJCAI,
2016.
[21] W. Yin, A. Zubiaga, Towards generalisable hate speech detection: a review on obstacles
and solutions, PeerJ. Computer science 7 (2021) e598–e598.
[22] J.-M. Xu, K.-S. Jun, X. Zhu, A. Bellmore, Learning from bullying traces in social
media, in: Proceedings of the 2012 Conference of the North American Chapter of
the Association for Computational Linguistics: Human Language Technologies,
Association for Computational Linguistics, Montréal, Canada, 2012, pp. 656–666. URL:
https://aclanthology.org/N12-1084.
[23] T. Davidson, D. Warmsley, M. W. Macy, I. Weber, Automated Hate Speech Detection and
the Problem of Ofensive Language, in: ICWSM, 2017.
[24] Y. Mehdad, J. Tetreault, Do characters abuse more than words?, in: Proceedings of the
17th Annual Meeting of the Special Interest Group on Discourse and Dialogue, Association
for Computational Linguistics, Los Angeles, 2016, pp. 299–303. URL: https://aclanthology.
org/W16-3638. doi:1 0 . 1 8 6 5 3 / v 1 / W 1 6 - 3 6 3 8 .
[25] J. S. Meyer, B. Gambäck, A platform agnostic dual-strand hate speech detector, in:
Proceedings of the Third Workshop on Abusive Language Online, Association for Computational
Linguistics, Florence, Italy, 2019, pp. 146–156. URL: https://aclanthology.org/W19-3516.
doi:1 0 . 1 8 6 5 3 / v 1 / W 1 9 - 3 5 1 6 .
[26] S. D. Swamy, A. Jamatia, B. Gambäck, Studying generalisability across abusive language
detection datasets, in: Proceedings of the 23rd Conference on Computational Natural
Language Learning (CoNLL), Association for Computational Linguistics, Hong Kong, China,
2019, pp. 940–950. URL: https://aclanthology.org/K19-1088. doi:1 0 . 1 8 6 5 3 / v 1 / K 1 9 - 1 0 8 8 .
[27] T. Mandl, S. Modha, P. Majumder, D. Patel, M. Dave, C. Mandalia, A. Patel, Overview
of the HASOC track at FIRE 2019: Hate Speech and Ofensive Content Identification
in Indo-European Languages, Proceedings of the 11th Forum for Information Retrieval
Evaluation (2019).
[28] T. Mandl, S. Modha, G. K. Shahi, A. K. Jaiswal, D. Nandini, D. Patel, P. Majumder, J. Schäfer,
Overview of the HASOC track at FIRE 2020: Hate Speech and Ofensive Content
Identification in Indo-European Languages, Proceedings of the 12th Forum for Information
Retrieval Evaluation (2020).
[29] I. Loshchilov, F. Hutter, Decoupled Weight Decay Regularization, in: ICLR, 2019.
[30] T. Wolf, L. Debut, V. Sanh, J. Chaumond, C. Delangue, A. Moi, P. Cistac, T. Rault, R. Louf,
M. Funtowicz, J. Davison, S. Shleifer, P. von Platen, C. Ma, Y. Jernite, J. Plu, C. Xu, T. L. Scao,
S. Gugger, M. Drame, Q. Lhoest, A. M. Rush, HuggingFace’s Transformers: State-of-the-art
Natural Language Processing, 2020. a r X i v : 1 9 1 0 . 0 3 7 7 1 .</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>O.</given-names>
            <surname>Istaiteh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Al-Omoush</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Tedmori</surname>
          </string-name>
          , Racist and Sexist Hate Speech Detection: Literature Review,
          <source>2020 International Conference on Intelligent Data Science Technologies and Applications (IDSTA)</source>
          (
          <year>2020</year>
          )
          <fpage>95</fpage>
          -
          <lpage>99</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>S.</given-names>
            <surname>Kawate</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Patil</surname>
          </string-name>
          ,
          <article-title>Analysis of foul language usage in social media text conversation</article-title>
          ,
          <source>Int. J. Soc. Media Interact. Learn. Environ</source>
          .
          <volume>5</volume>
          (
          <year>2017</year>
          )
          <fpage>227</fpage>
          -
          <lpage>251</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>S.</given-names>
            <surname>Jaki</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T. D.</given-names>
            <surname>Smedt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Gwóźdź</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Panchal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Rossa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G. D.</given-names>
            <surname>Pauw</surname>
          </string-name>
          ,
          <article-title>Online hatred of women in the Incels</article-title>
          .
          <source>me forum 7</source>
          (
          <year>2019</year>
          )
          <fpage>240</fpage>
          -
          <lpage>268</lpage>
          . URL: https://doi.org/10.1075%
          <fpage>2Fjlac</fpage>
          .00026.jak.
          <source>doi:1 0 . 1 0</source>
          <volume>7 5</volume>
          / j l a c .
          <article-title>0 0 0 2 6 . j a k .</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>B.</given-names>
            <surname>Mathew</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Dutt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Goyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mukherjee</surname>
          </string-name>
          ,
          <article-title>Spread of Hate Speech in Online Social Media</article-title>
          ,
          <source>Proceedings of the 10th ACM Conference on Web Science</source>
          (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>S.</given-names>
            <surname>Modha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Mandl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G. K.</given-names>
            <surname>Shahi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Madhu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Satapara</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Ranasinghe</surname>
          </string-name>
          ,
          <string-name>
            <surname>M.</surname>
          </string-name>
          <article-title>Zampieri, Overview of the HASOC Subtrack at FIRE 2021: Hate Speech and Ofensive Content Identification in English and Indo-Aryan Languages and Conversational Hate Speech</article-title>
          , in: FIRE 2021:
          <article-title>Forum for Information Retrieval Evaluation, Virtual Event</article-title>
          ,
          <fpage>13th</fpage>
          -17th
          <source>December</source>
          <year>2021</year>
          , ACM,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>T.</given-names>
            <surname>Mandl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Modha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G. K.</given-names>
            <surname>Shahi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Madhu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Satapara</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Majumder</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Schäfer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Ranasinghe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Zampieri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Nandini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. K.</given-names>
            <surname>Jaiswal</surname>
          </string-name>
          ,
          <article-title>Overview of the HASOC subtrack at FIRE 2021: Hate Speech and Ofensive Content Identification in English and Indo-Aryan Lan-</article-title>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>