<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>Forum for Information Retrieval Evaluation, December</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Fine-tuning of Pre-trained Transformers for Hate, Ofensive, and Profane Content Detection in English and Marathi</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Anna Glazkova</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Michael Kadantsev</string-name>
          <email>michael.kadantsev@gmail.com</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Maksim Glazkov</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>RoBERTa</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>LaBSE</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Marathi</string-name>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Neuro.net</institution>
          ,
          <addr-line>6/16 Alekseevskaya St, Nizhny Novgorod, 603000, Russian Federation</addr-line>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Thales Canada</institution>
          ,
          <addr-line>Transportation Solutions, 105 Moatfield Dr., Toronto</addr-line>
          ,
          <country country="CA">Canada</country>
          ,
          <addr-line>M3B 0A4</addr-line>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>University of Tyumen</institution>
          ,
          <addr-line>6 Volodarskogo St, Tyumen, 625003, Russian Federation</addr-line>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2021</year>
      </pub-date>
      <volume>1</volume>
      <fpage>3</fpage>
      <lpage>17</lpage>
      <abstract>
        <p>This paper describes neural models developed for the Hate Speech and Ofensive Content Identification in English and Indo-Aryan Languages Shared Task 2021. Our team called neuro-utmn-thales participated in two tasks on binary and fine-grained classification of English tweets that contain hate, ofensive, and profane content (English Subtasks A &amp; B) and one task on identification of problematic content in Marathi (Marathi Subtask A). For English subtasks, we investigate the impact of additional corpora for hate speech detection to fine-tune transformer models. We also apply a one-vs-rest approach based on Twitter-RoBERTa to discrimination between hate, profane and ofensive posts. Our models ranked third in English Subtask A with the F1-score of 81.99% and ranked second in English Subtask B with the F1-score of 65.77%. For the Marathi tasks, we propose a system based on the Language-Agnostic BERT Sentence Embedding (LaBSE). This model achieved the second result in Marathi Subtask A obtaining an media safer. The Hate Speech and Ofensive Content Identification in English and Indo-Aryan</p>
      </abstract>
      <kwd-group>
        <kwd>Hate speech</kwd>
        <kwd>ofensive language identification</kwd>
        <kwd>text classification</kwd>
        <kwd>transformer neural networks</kwd>
        <kwd>Twitter-</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Social media has a greater impact on our society. Social networks give us almost limitless
freedom of speech and contribute to the rapid dissemination of information. However, these
positive properties often lead to unhealthy usage of social media. Thus, hate speech spreading
afects users’ psychological state, promotes violence, and reinforces hateful sentiments [
        <xref ref-type="bibr" rid="ref1 ref2">1, 2</xref>
        ].
This problem attracts many scholars to apply modern technologies in order to make social
Languages Shared Task (HASOC) 2021 [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] aims to сompare and analyze existing approaches to
identifying hate speech not only for English, but also for other languages. It focused on detecting
hate, ofensive, and profane content in tweets, and ofering six subtasks. We participated in
three of them:
nEvelop-O
(M. Glazkov)
• English Subtask A: identifying hate, ofensive, and profane content from the post in
      </p>
      <p>
        English [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ].
• English Subtask B: discrimination between hate, profane, and ofensive posts in English.
• Marathi Subtask A: identifying hate, ofensive, and profane content from the post in
      </p>
      <p>
        Marathi [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ].
      </p>
      <p>The source code for our models is freely available1.</p>
      <p>The paper is organized as follows. Section 2 contains a brief review of related works. Next,
we describe our experiments on the binary and fine-grained classification of English tweets
in Section 3. In Section 4, we present our model for hate, ofensive, and profane language
identification in Marathi. We conclude this paper in Section 5. Finally, Section 6 contains
acknowledgments.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Works</title>
      <p>
        We briefly discuss works done related to harmful content detection in the past few years. Shared
tasks related to hate speech and ofensive language detection from tweets was organized as
a part of some workshops and conferences, such as FIRE [
        <xref ref-type="bibr" rid="ref6 ref7">6, 7</xref>
        ], SemEval, [
        <xref ref-type="bibr" rid="ref8 ref9">8, 9</xref>
        ], GermEval
[
        <xref ref-type="bibr" rid="ref10 ref11">10, 11</xref>
        ], IberLEF [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ], and OSACT [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ]. The participants proposed a broad range of approaches
from traditional machine learning techniques (for example, Support Vector Machines [
        <xref ref-type="bibr" rid="ref14 ref15">14, 15</xref>
        ],
Random Forest [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ]) to various neural architectures (Convolutional Neural Networks, CNN
[17]; Long Short Term Memory, LSTM [18, 19]; Embeddings from Language Models, ELMo [20];
and Bidirectional Encoder Representations from Transformers, BERT [21, 22]). In most cases,
BERT-based systems outperformed other approaches.
      </p>
      <p>
        Most research on hate speech detection continues to be based on English corpora. Despite this,
the harmful content is distributed in diferent languages. Therefore, there have been previous
attempts at creating corpora and developing models for hate speech detection in common
non-English languages, such as Arabic [
        <xref ref-type="bibr" rid="ref13">13, 23</xref>
        ], German [
        <xref ref-type="bibr" rid="ref10 ref11 ref6 ref7">6, 7, 10, 11</xref>
        ], Italian [24, 25], Spanish
[
        <xref ref-type="bibr" rid="ref12 ref9">9, 12</xref>
        ], Hindi [
        <xref ref-type="bibr" rid="ref6 ref7">6, 7</xref>
        ], Tamil and Malayalam [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. Several studies have focused on collecting hate
speech corpora for Chinese [26], Portuguese [27], Polish [28], Turkish [29] and Russian [30]
languages.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. English Subtasks A &amp; B: Identification and Fine-grained</title>
    </sec>
    <sec id="sec-4">
      <title>Classification of Hate, Ofensive, and Profane Tweets</title>
      <p>The objective of English Subtasks A &amp; B is to identify whether a tweet in English contains harmful
content (Subtask A) and perform a fine-grained classification of posts into three categories,
including: hate, ofensive, or profane (Subtask B).
3.1. Data
The dataset provided to the participants of the shared task contains 4355 manually annotated
social media posts divided into training (3074) and test (1281) sets. Table 1 presents the data
description.</p>
      <p>
        Further, we tested several data sampling techniques using diferent hate speech corpora as
additional training data. Firstly, we evaluated the joint use of multilingual data provided by
the organizers of HASOC 2021, including the English, the Hindi, and the Marathi training sets.
Secondly, as the training sets were highly imbalanced, we applied the positive class random
oversampling technique so that each training batch contained approximately the same number
of samples. Besides, we experimented with the seq2seq-based data augmentation technique
[31]. For this purpose, we fine-tuned the BART-base model for the denoising reconstruction
task where 40% of tokens are masked and the goal of the decoder is to reconstruct the original
sequence. Since the BART model [
        <xref ref-type="bibr" rid="ref17">32</xref>
        ] already contains the &lt;mask&gt; token, we use it to replace
mask tokens. We generated one synthetic example for every tweet in the training set. Thus, the
augmented data size is the same size as the size of the original training set. Finally, we evaluated
the impact of additional training data, including: (a) the English dataset, used at HASOC 2020 [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ];
(b) HatebaseTwitter, based on the hate speech lexicon from Hatebase2 [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]; (c) HatEval, a dataset
presented at Semeval-2019 Task 5 [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] ; (d) Ofensive Language Identification Dataset (OLID),
used in the SemEval-2019 Task 6 (OfensEval) [
        <xref ref-type="bibr" rid="ref18">33</xref>
        ]. All corpora except the HatebaseTwitter
dataset contain non-intersective classes. Besides, all listed datasets are collected from Twitter.
A representative sampling of additional data is shown in Table 2.
      </p>
      <p>
        We preprocessed the datasets for Subtasks A &amp; B in a similar manner. Inspired by [
        <xref ref-type="bibr" rid="ref19">34</xref>
        ], we
used the following text preprocessing technique3: (a) removed all URLs; (b) replaced all user
mentions with the $MENTION$ placeholder.
      </p>
      <p>2https://hatebase.org/
3https://pypi.org/project/tweet-preprocessor</p>
      <sec id="sec-4-1">
        <title>3.2. Models</title>
        <p>
          We conduct our experiments with neural models based on BERT [
          <xref ref-type="bibr" rid="ref20">35</xref>
          ] as they have achieved
state-of-the-art results in harmful content detection. For example, BERT-based models proved
eficient at previous HASOC shared tasks [
          <xref ref-type="bibr" rid="ref6 ref7">7, 6</xref>
          ] and SemEval [
          <xref ref-type="bibr" rid="ref18 ref21">33, 36</xref>
          ].
        </p>
        <p>
          We used the following models:
• BERT [
          <xref ref-type="bibr" rid="ref20">35</xref>
          ], a pre-trained model on BookCorpus [
          <xref ref-type="bibr" rid="ref22">37</xref>
          ] and English Wikipedia using a
masked language modeling objective.
• BERTweet [
          <xref ref-type="bibr" rid="ref23">38</xref>
          ], a pre-trained language model for English tweets. The corpus used to
pre-train BERTweet consists of 850M English Tweets including 845M Tweets streamed
from 01/2012 to 08/2019 and 5M Tweets related to the COVID-19 pandemic.
• Twitter-RoBERTa for Hate Speech Detection [
          <xref ref-type="bibr" rid="ref19">34</xref>
          ], a RoBERTa [
          <xref ref-type="bibr" rid="ref24">39</xref>
          ] model trained
on 58M tweets and fine-tuned for hate speech detection with the TweetEval benchmark.
• LaBSE [
          <xref ref-type="bibr" rid="ref25">40</xref>
          ], a language-agnostic BERT sentence embedding model supporting 109
languages.
        </p>
      </sec>
      <sec id="sec-4-2">
        <title>3.3. Experiments</title>
        <p>
          For both Subtask A and Subtask B, we adopted pre-trained models from HuggingFace [
          <xref ref-type="bibr" rid="ref26">41</xref>
          ] and
ifne-tuned them using PyTorch [
          <xref ref-type="bibr" rid="ref27">42</xref>
          ]. We fine-tuned each pre-trained language model for 3
epochs with the learning rate of 2e-5 using the AdamW optimizer [
          <xref ref-type="bibr" rid="ref28">43</xref>
          ]. We set batch size to 32
and maximum sequence size to 64. To validate our models during the development phase, we
divided labelled data using the train and validation split in the ratio 80:20.
        </p>
        <p>Table 3 shows the performance of our models on the validation subset for Subtask A in terms
of macro-averaging F1-score (F1), precision (P), and recall (R). As can be seen from the table,
BERT, BERTweet, and LaBSE show very close results during validation. Despite this, LaBSE
jointly fine-tuned on three mixed multilingual datasets shows the highest precision score. The
use of Twitter-RoBERTa increases the F1-score by 1.5-2.5% compared to other classification
models. Based on this, we chose Twitter-RoBERTa for further experiments. We found out
that neither the random oversampling technique nor the use of the augmented and additional
data shows a performance improvement, except the joint use of the original dataset and the
HatebaseTwitter dataset that gives an F1-score growth of 0.09% and a precision growth of 0.28%
compared to basic Twitter-RoBERTa.</p>
        <p>For our oficial submission for Subtask A, we designed a soft-voting ensemble of five
TwitterRoBERTa jointly fine-tuned on the original training set and the HatebaseTwitter dataset (see
Table 4). For Subtask B, we used the following one-vs-rest approach to discrimination between
hate, profane, and ofensive posts.</p>
        <p>• First, we applied our Subtask A binary models to identify non hate-ofensive examples.
• Second, we fine-tuned three Twitter-RoBERTa binary models to delimit examples of
hate-vs-profane, hate-vs-ofensive, and ofensive-vs-profane classes. The training dataset
was extended with the HatebaseTwitter dataset.
• Finally, we compared the results of binary models. If the result was defined uniquely, we
used it as a predicted label. Otherwise, we chose the label in proportion to the number of
examples in the training set.</p>
        <p>This can be illustrated briefly by the following examples.</p>
        <p>– Let the models show the following results:
∗ hate-vs-profane→hate;
∗ hate-vs-ofensive→hate;
∗ ofensive-vs-profane→ofensive.</p>
        <p>Thus, classes have the following votes: hate – 2, ofensive - 1, profane – 0. Then we
predict the HATE label.
– If the results are:
∗ hate-vs-profane→profane;
∗ hate-vs-ofensive→hate;
∗ ofensive-vs-profane→ofensive,
we have the class votes, such as hate – 1, ofensive - 1, profane – 1. Then we choose
the PRFN label as the most common label in the training set.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>4. Marathi Subtask A: Identifying Hate, Ofensive, and Profane</title>
    </sec>
    <sec id="sec-6">
      <title>Content from the Post</title>
      <p>
        4.1. Data
For the Marathi task, we used the original training and test sets provided by the organizers of
the HASOC 2021. The whole dataset contains 2499 tweets, including: 1874 training and 625 test
examples. The training set consists of 1205 texts of the NOT class and 669 texts of the HOF class.
We used raw data as an input for our models. Following [
        <xref ref-type="bibr" rid="ref29 ref30">44, 45</xref>
        ], we experimented with the
combination of the English, the Hindi, and the Marathi training sets provided by the organizers.
      </p>
      <sec id="sec-6-1">
        <title>4.2. Models</title>
        <p>
          We evaluated the following models:
• XLM-RoBERTa [
          <xref ref-type="bibr" rid="ref31">46</xref>
          ], a transformer-based multilingual masked language model
supporting 100 languages.
• LaBSE [
          <xref ref-type="bibr" rid="ref25">40</xref>
          ], a language-agnostic BERT sentence embedding model pre-trained on texts in
109 languages.
        </p>
      </sec>
      <sec id="sec-6-2">
        <title>4.3. Experiments</title>
        <p>We experimented with the above-mentioned language models fine-tuned on monolingual and
multilingual data. For model evaluation during the development phase, we used the random
train and validation split in the ratio 80:20 with a fixed seed. We set the same model parameters
as for English tasks.</p>
        <p>Table 5 illustrates the results. It can be seen that LaBSE outperforms XLM-RoBERTa in all
cases. Moreover, the F1-score of LaBSE fine-tuned only on the Marathi dataset are higher than
the results of LaBSE fine-tuned on multilingual data. XLM-RoBERTa, on the other hand, mostly
benefits from multilingual fine-tuning.</p>
        <p>For our final submission, we used a soft-voting ensemble of five LaBSE fine-tuned on the
oficial Marathi dataset provided by the organizers of the competition. The results of this model
on the test set are shown in Table 6.</p>
      </sec>
    </sec>
    <sec id="sec-7">
      <title>5. Conclusion</title>
      <p>In this paper, we have presented the details about our participation in the HASOC Shared
Task 2021. We have explored an application of domain-specific monolingual and multilingual
BERT-based models to the tasks of binary and fine-grained classification of Twitter posts. We
also proposed a one-vs-rest approach to discrimination between hate, ofensive, and profane
tweets. Further research can focus on analyzing the efectiveness of various text
preprocessing techniques for harmful content detection and exploring how diferent transfer learning
approaches can afect classification performance.</p>
    </sec>
    <sec id="sec-8">
      <title>6. Acknowledgments</title>
      <p>The work on multi-label text classification was carried out by Anna Glazkova and supported by
the grant of the President of the Russian Federation no. MK-637.2020.9.
hate speech and ofensive content identification in Indo-European languages, in: FIRE
(Working Notes), 2020, pp. 168–174.
[17] A. Ribeiro, N. Silva, Inf-hateval at semeval-2019 task 5: Convolutional neural networks
for hate speech detection against women and immigrants on Twitter, in: Proceedings of
the 13th International Workshop on Semantic Evaluation, 2019, pp. 420–425.
[18] A. K. Mishraa, S. Saumyab, A. Kumara, IIIT_DWD@ HASOC 2020: Identifying ofensive
content in Indo-European languages (2020).
[19] A. Montejo-Ráez, S. M. Jiménez-Zafra, M. A. García-Cumbreras, M. C. Díaz-Galiano,
SINAIDL at SemEval-2019 task 5: Recurrent networks and data augmentation by paraphrasing,
in: Proceedings of the 13th International Workshop on Semantic Evaluation, 2019, pp.
480–483.
[20] M. Bojkovsky, M. Pikuliak, STUFIIT at SemEval-2019 task 5: Multilingual hate speech
detection on twitter with MUSE and ELMo embeddings, in: Proceedings of the 13th
International Workshop on Semantic Evaluation, 2019, pp. 464–468.
[21] J. Risch, A. Stoll, M. Ziegele, R. Krestel, hpiDEDIS at GermEval 2019: Ofensive language
identification using a German BERT model., in: KONVENS, 2019.
[22] P. Liu, W. Li, L. Zou, NULI at SemEval-2019 task 6: Transfer learning for ofensive language
detection using bidirectional transformers, in: Proceedings of the 13th international
workshop on semantic evaluation, 2019, pp. 87–91.
[23] N. Albadi, M. Kurdi, S. Mishra, Are they our brothers? analysis and detection of religious
hate speech in the Arabic twittersphere, in: 2018 IEEE/ACM International Conference on
Advances in Social Networks Analysis and Mining (ASONAM), IEEE, 2018, pp. 69–76.
[24] C. Bosco, D. Felice, F. Poletto, M. Sanguinetti, T. Maurizio, Overview of the EVALITA
2018 hate speech detection task, in: EVALITA 2018-Sixth Evaluation Campaign of Natural
Language Processing and Speech Tools for Italian, volume 2263, CEUR, 2018, pp. 1–9.
[25] M. Sanguinetti, F. Poletto, C. Bosco, V. Patti, M. Stranisci, An Italian Twitter corpus of hate
speech against immigrants, in: Proceedings of the Eleventh International Conference on
Language Resources and Evaluation (LREC 2018), 2018.
[26] X. Tang, X. Shen, Y. Wang, Y. Yang, Categorizing ofensive language in social networks:
A Chinese corpus, systems and an explanation tool, in: China National Conference on
Chinese Computational Linguistics, Springer, 2020, pp. 300–315.
[27] R. P. de Pelle, V. P. Moreira, Ofensive comments in the Brazilian web: a dataset and
baseline results, in: Anais do VI Brazilian Workshop on Social Network Analysis and
Mining, SBC, 2017.
[28] M. Ptaszynski, A. Pieciukiewicz, P. Dybała, Results of the PolEval 2019 shared task 6:
First dataset and open shared task for automatic cyberbullying detection in Polish Twitter
(2019).
[29] Ç. Çöltekin, A corpus of Turkish ofensive language on social media, in: Proceedings of
the 12th Language Resources and Evaluation Conference, 2020, pp. 6174–6184.
[30] L. Komalova, A. Glazkova, D. Morozov, R. Epifanov, L. Motovskikh, E. Mayorova,
Automated classification of potentially insulting speech acts on social network sites, in:
International Conference on Digital Transformation and Global Society, Springer, 2021.
[31] V. Kumar, A. Choudhary, E. Cho, Data augmentation using pre-trained transformer models,
in: Proceedings of the 2nd Workshop on Life-long Learning for Spoken Language Systems,
M. Ott, L. Zettlemoyer, V. Stoyanov, Unsupervised cross-lingual representation learning at
scale, in: Proceedings of the 58th Annual Meeting of the Association for Computational
Linguistics, 2020, pp. 8440–8451.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>L. E.</given-names>
            <surname>Beausoleil</surname>
          </string-name>
          , Free, hateful, and
          <article-title>posted: rethinking first amendment protection of hate speech in a social media world</article-title>
          ,
          <source>BCL Rev</source>
          .
          <volume>60</volume>
          (
          <year>2019</year>
          )
          <fpage>2101</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>M.</given-names>
            <surname>Bilewicz</surname>
          </string-name>
          , W. Soral,
          <article-title>Hate speech epidemic. the dynamic efects of derogatory language on intergroup relations and political radicalization</article-title>
          ,
          <source>Political Psychology</source>
          <volume>41</volume>
          (
          <year>2020</year>
          )
          <fpage>3</fpage>
          -
          <lpage>33</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>S.</given-names>
            <surname>Modha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Mandl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G. K.</given-names>
            <surname>Shahi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Madhu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Satapara</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Ranasinghe</surname>
          </string-name>
          ,
          <string-name>
            <surname>M.</surname>
          </string-name>
          <article-title>Zampieri, Overview of the HASOC subtrack at FIRE 2021: Hate speech and ofensive content identiifcation in English and Indo-Aryan languages and conversational hate speech</article-title>
          ,
          <source>in: FIRE</source>
          <year>2021</year>
          :
          <article-title>Forum for Information Retrieval Evaluation, Virtual Event</article-title>
          ,
          <fpage>13th</fpage>
          -17th
          <source>December</source>
          <year>2021</year>
          , ACM,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>T.</given-names>
            <surname>Mandl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Modha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G. K.</given-names>
            <surname>Shahi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Madhu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Satapara</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Majumder</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Schäfer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Ranasinghe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Zampieri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Nandini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. K.</given-names>
            <surname>Jaiswal</surname>
          </string-name>
          ,
          <source>Overview of the HASOC subtrack at FIRE</source>
          <year>2021</year>
          :
          <article-title>Hate speech and ofensive content identification in English and Indo-Aryan languages</article-title>
          , in: Working Notes of FIRE 2021 -
          <article-title>Forum for Information Retrieval Evaluation</article-title>
          ,
          <string-name>
            <surname>CEUR</surname>
          </string-name>
          ,
          <year>2021</year>
          . URL: http://ceur-ws.org/.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>S.</given-names>
            <surname>Gaikwad</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Ranasinghe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Zampieri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. M.</given-names>
            <surname>Homan</surname>
          </string-name>
          ,
          <article-title>Cross-lingual ofensive language identification for low resource languages: The case of Marathi</article-title>
          ,
          <source>in: Proceedings of RANLP</source>
          ,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>T.</given-names>
            <surname>Mandl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Modha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Majumder</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Patel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Dave</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Mandlia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Patel</surname>
          </string-name>
          ,
          <source>Overview of the HASOC track at FIRE</source>
          <year>2019</year>
          :
          <article-title>Hate speech and ofensive content identification in Indo-European languages</article-title>
          ,
          <source>in: Proceedings of the 11th forum for information retrieval evaluation</source>
          ,
          <year>2019</year>
          , pp.
          <fpage>14</fpage>
          -
          <lpage>17</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>T.</given-names>
            <surname>Mandl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Modha</surname>
          </string-name>
          ,
          <string-name>
            <surname>A. Kumar</surname>
            <given-names>M</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>B. R.</given-names>
            <surname>Chakravarthi</surname>
          </string-name>
          ,
          <source>Overview of the HASOC track at FIRE</source>
          <year>2020</year>
          :
          <article-title>Hate speech and ofensive language identification in Tamil, Malayalam, Hindi, English and German</article-title>
          ,
          <source>in: Forum for Information Retrieval Evaluation</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>29</fpage>
          -
          <lpage>32</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>T.</given-names>
            <surname>Davidson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Warmsley</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Macy</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Weber</surname>
          </string-name>
          ,
          <article-title>Automated hate speech detection and the problem of ofensive language</article-title>
          ,
          <source>in: Proceedings of the International AAAI Conference on Web and Social Media</source>
          , volume
          <volume>11</volume>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>V.</given-names>
            <surname>Basile</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Bosco</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Fersini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Debora</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Patti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F. M. R.</given-names>
            <surname>Pardo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Rosso</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Sanguinetti</surname>
          </string-name>
          , et al.,
          <article-title>Semeval-2019 task 5: Multilingual detection of hate speech against immigrants and women in Twitter</article-title>
          ,
          <source>in: 13th International Workshop on Semantic Evaluation</source>
          ,
          <year>2019</year>
          , pp.
          <fpage>54</fpage>
          -
          <lpage>63</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>M.</given-names>
            <surname>Wiegand</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Siegel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Ruppenhofer</surname>
          </string-name>
          ,
          <article-title>Overview of the germeval 2018 shared task on the identification of ofensive language</article-title>
          ,
          <source>in: 14th Conference on Natural Language Processing KONVENS</source>
          <year>2018</year>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>J. M. Struß</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Siegel</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Ruppenhofer</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Wiegand</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Klenner</surname>
          </string-name>
          , et al.,
          <source>Overview of GermEval task 2</source>
          ,
          <year>2019</year>
          <article-title>shared task on the identification of ofensive language (</article-title>
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>M.</given-names>
            <surname>Taulé</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Ariza</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Nofre</surname>
          </string-name>
          , E. Amigó,
          <string-name>
            <given-names>P.</given-names>
            <surname>Rosso</surname>
          </string-name>
          , Overview of detoxis at IberLEF 2021:
          <article-title>Detection of toxicity in comments in Spanish</article-title>
          ,
          <source>Procesamiento del Lenguaje Natural</source>
          <volume>67</volume>
          (
          <year>2021</year>
          )
          <fpage>209</fpage>
          -
          <lpage>221</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>H.</given-names>
            <surname>Mubarak</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Darwish</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Magdy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Elsayed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Al-Khalifa</surname>
          </string-name>
          ,
          <article-title>Overview of OSACT4 Arabic ofensive language detection shared task</article-title>
          ,
          <source>in: Proceedings of the 4th Workshop on Open-Source Arabic Corpora and Processing Tools, with a Shared Task on Ofensive Language Detection</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>48</fpage>
          -
          <lpage>52</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>F.</given-names>
            <surname>Schmid</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Thielemann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mantwill</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Xi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Labudde</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Spranger</surname>
          </string-name>
          ,
          <article-title>Fosil-ofensive language classification of German tweets combining SVMs and deep learning techniques</article-title>
          ,
          <source>in: KONVENS</source>
          ,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>S.</given-names>
            <surname>Hassan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Samih</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Mubarak</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Abdelali</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Rashed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. A.</given-names>
            <surname>Chowdhury</surname>
          </string-name>
          ,
          <article-title>Alt submission for osact shared task on ofensive language detection</article-title>
          ,
          <source>in: Proceedings of the 4th Workshop on Open-Source Arabic Corpora and Processing Tools, with a Shared Task on Ofensive Language Detection</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>61</fpage>
          -
          <lpage>65</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>B.</given-names>
            <surname>Ray</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Garain</surname>
          </string-name>
          , JU at HASOC 2020:
          <article-title>Deep learning with RoBERTa and random forest for 2020</article-title>
          , pp.
          <fpage>18</fpage>
          -
          <lpage>26</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [32]
          <string-name>
            <given-names>M.</given-names>
            <surname>Lewis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Goyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ghazvininejad</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mohamed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Levy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Stoyanov</surname>
          </string-name>
          , L. Zettlemoyer, Bart:
          <article-title>Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension</article-title>
          ,
          <source>in: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>7871</fpage>
          -
          <lpage>7880</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [33]
          <string-name>
            <given-names>M.</given-names>
            <surname>Zampieri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Malmasi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Nakov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Rosenthal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Farra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Kumar</surname>
          </string-name>
          , Semeval
          <article-title>-2019 task 6: Identifying and categorizing ofensive language in social media (OfensEval)</article-title>
          ,
          <source>in: Proceedings of the 13th International Workshop on Semantic Evaluation</source>
          ,
          <year>2019</year>
          , pp.
          <fpage>75</fpage>
          -
          <lpage>86</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [34]
          <string-name>
            <given-names>F.</given-names>
            <surname>Barbieri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Camacho-Collados</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. E.</given-names>
            <surname>Anke</surname>
          </string-name>
          , L. Neves,
          <article-title>TweetEval: Unified benchmark and comparative evaluation for tweet classification</article-title>
          ,
          <source>in: Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: Findings</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>1644</fpage>
          -
          <lpage>1650</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [35]
          <string-name>
            <given-names>J.</given-names>
            <surname>Devlin</surname>
          </string-name>
          , M.-
          <string-name>
            <given-names>W.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Toutanova</surname>
          </string-name>
          , Bert:
          <article-title>Pre-training of deep bidirectional transformers for language understanding</article-title>
          , arXiv preprint arXiv:
          <year>1810</year>
          .
          <volume>04805</volume>
          (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [36]
          <string-name>
            <given-names>J.</given-names>
            <surname>Pavlopoulos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Sorensen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Laugier</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Androutsopoulos</surname>
          </string-name>
          , SemEval
          <article-title>-2021 task 5: Toxic spans detection</article-title>
          ,
          <source>in: Proceedings of the 15th International Workshop on Semantic Evaluation (SemEval-2021)</source>
          ,
          <year>2021</year>
          , pp.
          <fpage>59</fpage>
          -
          <lpage>69</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [37]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Kiros</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Zemel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Salakhutdinov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Urtasun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Torralba</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Fidler</surname>
          </string-name>
          ,
          <article-title>Aligning books and movies: Towards story-like visual explanations by watching movies and reading books</article-title>
          ,
          <source>in: Proceedings of the IEEE international conference on computer vision</source>
          ,
          <year>2015</year>
          , pp.
          <fpage>19</fpage>
          -
          <lpage>27</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [38]
          <string-name>
            <given-names>D. Q.</given-names>
            <surname>Nguyen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Vu</surname>
          </string-name>
          , A. T. Nguyen,
          <article-title>BERTweet: A pre-trained language model for English tweets</article-title>
          ,
          <source>in: Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>9</fpage>
          -
          <lpage>14</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [39]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ott</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Goyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Du</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Joshi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Levy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Lewis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Zettlemoyer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Stoyanov</surname>
          </string-name>
          ,
          <article-title>Roberta: A robustly optimized bert pretraining approach</article-title>
          , arXiv preprint arXiv:
          <year>1907</year>
          .
          <volume>11692</volume>
          (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [40]
          <string-name>
            <given-names>F.</given-names>
            <surname>Feng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Cer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Arivazhagan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <article-title>Language-agnostic BERT sentence embedding</article-title>
          , arXiv preprint arXiv:
          <year>2007</year>
          .
          <year>01852</year>
          (
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [41]
          <string-name>
            <given-names>T.</given-names>
            <surname>Wolf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Chaumond</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Debut</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Sanh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Delangue</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Moi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Cistac</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Funtowicz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Davison</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Shleifer</surname>
          </string-name>
          , et al.,
          <article-title>Transformers: State-of-the-art natural language processing</article-title>
          ,
          <source>in: Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>38</fpage>
          -
          <lpage>45</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [42]
          <string-name>
            <given-names>A.</given-names>
            <surname>Paszke</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Gross</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Massa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Lerer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Bradbury</surname>
          </string-name>
          , G. Chanan,
          <string-name>
            <given-names>T.</given-names>
            <surname>Killeen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Gimelshein</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Antiga</surname>
          </string-name>
          , et al.,
          <article-title>Pytorch: An imperative style, high-performance deep learning library</article-title>
          ,
          <source>Advances in neural information processing systems</source>
          <volume>32</volume>
          (
          <year>2019</year>
          )
          <fpage>8026</fpage>
          -
          <lpage>8037</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          [43]
          <string-name>
            <given-names>I.</given-names>
            <surname>Loshchilov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Hutter</surname>
          </string-name>
          ,
          <article-title>Decoupled weight decay regularization</article-title>
          ,
          <source>in: International Conference on Learning Representations</source>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          [44]
          <string-name>
            <given-names>S.</given-names>
            <surname>Mishra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Prasad</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Mishra</surname>
          </string-name>
          ,
          <article-title>Multilingual joint fine-tuning of transformer models for identifying trolling</article-title>
          ,
          <source>aggression and cyberbullying at TRAC</source>
          <year>2020</year>
          , in: Proceedings of the Second Workshop on Trolling,
          <source>Aggression and Cyberbullying</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>120</fpage>
          -
          <lpage>125</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          [45]
          <string-name>
            <given-names>P.</given-names>
            <surname>Singh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Bhattacharyya</surname>
          </string-name>
          , CFILT IIT Bombay at HASOC 2020:
          <article-title>Joint multitask learning of multilingual hate speech and ofensive content detection system</article-title>
          .,
          <source>in: FIRE (Working Notes)</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>325</fpage>
          -
          <lpage>330</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          [46]
          <string-name>
            <given-names>A.</given-names>
            <surname>Conneau</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Khandelwal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Goyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Chaudhary</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Wenzek</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Guzmán</surname>
          </string-name>
          , É. Grave,
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>