<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>By1510 @ HaSpeeDe 2: Identification of Hate Speech for Italian Language in Social Media Data</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Tao Deng</string-name>
          <email>Dtao.top@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Yang Bai</string-name>
          <email>baiyang.top@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Hongbing Daiy</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>School of Information Science and Engineering Yunnan University</institution>
          ,
          <addr-line>Yunnan</addr-line>
          ,
          <country>P.R. China hbdai</country>
        </aff>
      </contrib-group>
      <abstract>
        <p />
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1 Introduction</title>
      <p>English. Hate speech detection has
become a crucial mission in many fields.
This paper introduces the system of team
By1510. In this work, we participate in
the HaSpeeDe 2 (Hate Speech Detection)
shared task which is organized within
Evalita 2020(The Final Workshop of the 7th
evaluation campaign). In order to obtain
more abundant semantic information, we
combine the original output of BERT-Ita
and the hidden state outputs of BERT-Ita.
We take part in task A. Our model achieves
an F1 score of 77.66% (6/27) in the tweets
test set and our model achieves an F1 score
of 66.38% (14/27) in the news headlines
test set.</p>
      <p>Italiano. L’ individuazione dell’
incitamento allodio diventata una
missione cruciale in molti campi. Questo
articolo introduce il sistema del team
By1510. In questo lavoro, partecipiamo
al task HaSpeeDe 2 che stato
organizzato allinterno di Evalita 2020. Per
ottenere informazioni semantiche pi
abbondanti abbiamo combinato loutput
originale di BERT Ita e gli output di hidden
state di BERT Ita. Il sistema presentato
partecipa al task A. Il nostro modello
ottiene un punteggio F1 di 77.66% (6/27) sui
dati di test da Twitter e un punteggio F1 di
66.38% (14/27) sui dati di test contenenti
titoli di quotidiano.</p>
      <p>Copyright c 2020 for this paper by its authors. Use
permitted under Creative Commons License Attribution 4.0
International (CC BY 4.0).</p>
      <p>
        With the continuous development of computer and
networks, social media users have increased year
by year, social media has entered people’s daily
life and becomes an indispensable part. More and
more people use the Internet to express their
opinions and ideas on social media platforms. Some
offensive, abusive, defamatory contents are easy to
spread and incite hatred, and these negative
contents can cause some bad effects. The simplest
way is that people mark the report and then delete
the system warning, which can not be solved
efficiently. Therefore, an efficient way is urgently
needed to eliminate these negative effects. This
paper proposes a hate speech detection system,
which can better detect and mark these
annoying contents. The HaSpeeDe 2
        <xref ref-type="bibr" rid="ref17">(Sanguinetti et al.,
2020)</xref>
        (Hate Speech Detection) shared task is
organized within Evalita 2020
        <xref ref-type="bibr" rid="ref3">(Basile et al., 2020)</xref>
        , the
7th evaluation campaign of Natural Language
Processing and Speech tools for Italian, which help
to detect whether the Italian language on
Twitter contains hate language, with the aim to reduce
the spread of hate speeches and online harassment.
(Waseem and Hovy, 2016)
      </p>
      <p>In this paper, we take part in task A in the
HaSpeeDe 2 task. The BERT model we use is
dbmz1 trained on Italian data. In order to obtain
more abundant semantic information, we
extract the state of hidden layer outputs and we provide
a reference for the detection of the hate speech in
the Italian language. The rest of the paper is
organized as follows. Section 2 briefly shows the
related work for the identification of hate speeches.
Section 3 elaborates on our approach. It shows the
data set officially provided and architecture of our
model. Section 4 describes the hyper-parameters
and our results. Finally, Section 5 concludes our
work.</p>
      <p>1https://huggingface.co/dbmdz</p>
    </sec>
    <sec id="sec-2">
      <title>Related Work</title>
      <p>
        Previously, machine learning
        <xref ref-type="bibr" rid="ref11 ref12 ref6">(Davidson et al.,
2017; MacAvaney et al., 2019a)</xref>
        , Bayesian method
        <xref ref-type="bibr" rid="ref1 ref13 ref16 ref20 ref9">(Miok et al., 2020; Fauzi and Yuniarti, 2018)</xref>
        ,
support vector machine
        <xref ref-type="bibr" rid="ref11 ref12 ref7">(MacAvaney et al., 2019b;
Del Vigna12 et al., 2017)</xref>
        , neural network
        <xref ref-type="bibr" rid="ref2 ref20">(Badjatiya et al., 2017; Zhang et al., 2018)</xref>
        and
other methods were proposed for the identification
of hate speech. In the Hindi-English mixed
language,
        <xref ref-type="bibr" rid="ref4">(Bohra et al., 2018)</xref>
        et al. in parentheses
used a supervised classification system to detect
the hate speech in the text in the code-mixed
language. The classification system used Character
N-Grams, Word N-Grams, Punctuations, Negation
Words, Lexicon and other feature vectors for
classification and training. The accuracy could reach
71.7% with SVM, which proved to be a very
effective method for classification tasks. In Danish
language,
        <xref ref-type="bibr" rid="ref10 ref18">(Sigurbergsson and Derczynski, 2019)</xref>
        developed four automatic classification systems
to detect and classify hate speech in English and
Danish, and proposed a method to automatically
detect different types of the hate speech, which
achieved good results for the detection of English
and Danish hate speeches. In English language,
        <xref ref-type="bibr" rid="ref1 ref16 ref20 ref9">(Aroyehun and Gelbukh, 2018)</xref>
        used a linear
baseline classifier (nbsvm with n-grams) and improved
deep neural network model.
      </p>
      <p>
        For the Italian language,
        <xref ref-type="bibr" rid="ref15">(Polignano et al.,
2019)</xref>
        proposed an AlBERT o model based on
classifier integration, which was verified by cross
validation on Facebook and Twitter data sets, and
the effect was obvious in offensive words.
        <xref ref-type="bibr" rid="ref5">(Corazza et al., 2018)</xref>
        used recurrent neural network,
ngram neural network and support vector machine
to classify Twitter data sets, and its recurrent
model had achieved good results. (Bianchini et al.,
2018) proposed artificial neural network to
annotate and classify 3000 message data from
Facebook and Twitter, and achieved good results.
3
3.1
      </p>
    </sec>
    <sec id="sec-3">
      <title>Methodology</title>
      <sec id="sec-3-1">
        <title>Data Description</title>
        <p>In this work, we take part in task A, which is a
binary classification task aimed at determining the
presence or the absence of hateful content in the
text towards a given target (among Immigrants,
Muslims or Roma people). The organizers
provide the training set and test set. For the training
set, it is from Twitter. For the test set, the
organizers provide in-domain data and out-of-domain
data, which come from Twitter and news headlines,
respectively. It can be seen from Table 1 that the
data set is slightly imbalanced.
3.2</p>
      </sec>
      <sec id="sec-3-2">
        <title>Our approach</title>
        <p>
          As the train data is very limited we resort to a
transfer learning approach. That is, we take an
NLP model pre-trained
          <xref ref-type="bibr" rid="ref14 ref16 ref8">(Peters et al., 2018;
Radford et al., 2018; Devlin et al., 2019)</xref>
          on a large
train data
test data
(tweets)
test data
(news headlines)
        </p>
        <p>Hate Speech
(HS)
2766
622
181</p>
        <p>No HS
4071
641
319
corpus of texts and fine-tune it for a specific task
at hand. In this work, we used
BERT-base-Italianuncased(BERT-Ita)2 from Transformers library. It
is trained on the recent Wikipedia dump and
various texts from the OPUS corpora3 collection. The
final training corpus has a size of 13GB and 2050
million tokens. For classification tasks, the
output of BERT-Ita (pooler output) is obtained by its
last layer hidden state of the first token of the
sequence (CLS token) further processed by a linear
layer and a Tanh activation function. However, the
pooler output is usually not a good summary of the
semantic information. Therefore, we extract the
hidden layer output of BERT-Ita to obtain more
abundant semantic information.</p>
        <p>
          <xref ref-type="bibr" rid="ref10">(Jawahar et al., 2019)</xref>
          pointed that the hidden
layer of BERT encodes a rich hierarchy of
linguistic information, with surface features at the
bottom layer, syntactic features in the middle layer
and semantic features at the top layer. Therefore,
we get abundant semantic information by
extracting the extra semantic features which is the last
three hidden layer outputs(L12 H0, L11 H0 and
L10 H0) of BERT-Ita. We propose the following
model which is shown in Figure 1. In the
model, we get L12 H0, L11 H0, L10 H0 from the top
2https://huggingface.co/dbmdz/bert-base-italian-uncased
3http://opus.nlpl.eu/
hidden layer of BERT-Ita. We concatenate pooler
output, L12 H0, L11 H0 and L10 H0 into the
classifier.
        </p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Experiments and Results</title>
      <sec id="sec-4-1">
        <title>Preprocessing and Experiments Setup</title>
        <p>In the experiment, we try to preprocess the
text but we did not achieve the desired results. We
find that after preprocessing the Twitter data, the
F1-score of the model decreased on the validation
set. We do not preprocess the data and we do not
use an extra data set. In this work, the training
set is split into the new training set and the
validation set by using the Stratified 5-Fold
Crossvalidation4.The random seed is set 42 in
Crossvalidation. Due to the imbalance of datasets, the
Stratified 5-Fold Cross-validation ensures that the
proportion of samples in each category in each
fold data set remains unchanged. During the
training, the best weight of the model is saved in 8
epochs. Table 2 shows the hyperparameters used in
our model.
4.2</p>
      </sec>
      <sec id="sec-4-2">
        <title>Results and analysis</title>
        <p>In the experiment, we find that with the increase
of the extra semantic features, the model can
obtain more abundant semantic information. Table 3
shows the performance of the model for different
semantic features after getting the labels of the test
set.5.</p>
        <p>No HS</p>
        <p>HS
No HS</p>
        <p>Hs</p>
        <sec id="sec-4-2-1">
          <title>Task A</title>
          <p>test set of tweets(100%)
No HS HS
489 152
119 503</p>
        </sec>
        <sec id="sec-4-2-2">
          <title>Task A</title>
          <p>test set of news headlines(100%)
No HS HS
312 7
133 48</p>
          <p>4https://scikit-learn.org/stable/modules/generated/sklearn
.model selection.StratifiedKFold.html#sklearn.model selecti
on.StratifiedKFold
5https://github.com/msang/haspeede/tree/master/2020
sults on the tweets test set, but the results of our
model are not good on the news headline data set.
There are many differences between the syntactic
features of tweets and news headlines. For
example, there are many irregular expressions in
tweets, while news expressions are very standard. Our
model is only fine-tuned on the tweets data set, so
we think this affects the performance of the model
on other types of data.
5</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Conclusion</title>
      <p>In this work, this paper introduces the system
proposed for HaSpeeDe 2 shared task for identifying
and classifying hate speeches on social media. We
enriched BERT-Ita with semantic information by
extracting the extra semantic features. We find
that with the increase of semantic information, the
performance of the model for identifying the hate
speech is also increasing. Finally, in the official
evaluation, our model rank 6th (6/27) in the tweets
test set and 14th (14/27) in the news headlines test
set. In the future, we will focus on how to make
the model learns more semantic information.</p>
      <p>The confusion matrices (actual values are
represented by rows) are shown in Table 4, Table 5,
Table 6. These tables show the performance of the
model on the test set as the extra semantic features
increase. In the tweets test set, we can see from
these tables that the ability of the model to detect
the hate speech is increasing as the extra
semantic features increase. Similarly, in the news
headlines test set, the ability of the model to detect the
hate speech is also increasing. We think that with
the increase of these extra semantic features, the
model can learn more semantic information. In
addition, we find that our model achieve good
reuation Campaign of Natural Language Processing
and Speech Tools for Italian. Final Workshop
(EVALITA 2020), Online. CEUR.org.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <string-name>
            <given-names>Segun</given-names>
            <surname>Taofeek</surname>
          </string-name>
          Aroyehun and
          <string-name>
            <given-names>Alexander</given-names>
            <surname>Gelbukh</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>Aggression detection in social media: Using deep neural networks, data augmentation, and pseudo labeling</article-title>
          .
          <source>In Proceedings of the First Workshop on Trolling, Aggression and Cyberbullying (TRAC2018)</source>
          , pages
          <fpage>90</fpage>
          -
          <lpage>97</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          <string-name>
            <given-names>Pinkesh</given-names>
            <surname>Badjatiya</surname>
          </string-name>
          , Shashank Gupta, Manish Gupta, and
          <string-name>
            <given-names>Vasudeva</given-names>
            <surname>Varma</surname>
          </string-name>
          .
          <year>2017</year>
          .
          <article-title>Deep learning for hate speech detection in tweets</article-title>
          .
          <source>In Proceedings of the 26th International Conference on World Wide Web Companion</source>
          , pages
          <fpage>759</fpage>
          -
          <lpage>760</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          <string-name>
            <given-names>Valerio</given-names>
            <surname>Basile</surname>
          </string-name>
          , Danilo Croce, Maria Di Maro, and
          <string-name>
            <surname>Lucia</surname>
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Passaro</surname>
          </string-name>
          .
          <year>2020</year>
          .
          <article-title>Evalita 2020: Overview of the 7th evaluation campaign of natural language processing and speech tools for italian</article-title>
          .
          <source>In Valerio Basile</source>
          , Danilo Croce, Maria Di Maro, and Lucia C. Passaro, editors,
          <source>Proceedings of Seventh EvalGiulio Bianchini</source>
          , Lore nzo Ferri, and
          <string-name>
            <given-names>Tommaso</given-names>
            <surname>Giorni</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>Text analysis for hate speech detection in italian messages on twitter and facebook</article-title>
          .
          <source>EVALITA Evaluation of NLP and Speech Tools for Italian</source>
          ,
          <volume>12</volume>
          :
          <fpage>250</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          <string-name>
            <given-names>Aditya</given-names>
            <surname>Bohra</surname>
          </string-name>
          , Deepanshu Vijay, Vinay Singh, Syed Sarfaraz Akhtar, and
          <string-name>
            <given-names>Manish</given-names>
            <surname>Shrivastava</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>A dataset of hindi-english code-mixed social media text for hate speech detection</article-title>
          .
          <source>In Proceedings of the second workshop on computational modeling of peoples opinions, personality, and emotions in social media</source>
          , pages
          <fpage>36</fpage>
          -
          <lpage>41</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          <string-name>
            <given-names>Michele</given-names>
            <surname>Corazza</surname>
          </string-name>
          , Stefano Menini, Pinar Arslan, Rachele Sprugnoli, Elena Cabrio, Sara Tonelli, and
          <string-name>
            <given-names>Serena</given-names>
            <surname>Villata</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>Comparing different supervised approaches to hate speech detection</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          <string-name>
            <given-names>Thomas</given-names>
            <surname>Davidson</surname>
          </string-name>
          , Dana Warmsley,
          <string-name>
            <given-names>Michael</given-names>
            <surname>Macy</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Ingmar</given-names>
            <surname>Weber</surname>
          </string-name>
          .
          <year>2017</year>
          .
          <article-title>Automated hate speech detection and the problem of offensive language</article-title>
          .
          <source>arXiv preprint arXiv:1703</source>
          .
          <fpage>04009</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          <string-name>
            <surname>Fabio Del Vigna12</surname>
          </string-name>
          ,
          <string-name>
            <surname>Andrea</surname>
            <given-names>Cimino23</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Felice</surname>
            <given-names>DellOrletta</given-names>
          </string-name>
          , Marinella Petrocchi, and
          <string-name>
            <given-names>Maurizio</given-names>
            <surname>Tesconi</surname>
          </string-name>
          .
          <year>2017</year>
          .
          <article-title>Hate me, hate me not: Hate speech detection on facebook</article-title>
          .
          <source>In Proceedings of the First Italian Conference on Cybersecurity (ITASEC17)</source>
          , pages
          <fpage>86</fpage>
          -
          <lpage>95</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          <string-name>
            <given-names>Jacob</given-names>
            <surname>Devlin</surname>
          </string-name>
          ,
          <string-name>
            <surname>Ming-Wei</surname>
            <given-names>Chang</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>Kenton</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>and Kristina</given-names>
            <surname>Toutanova</surname>
          </string-name>
          .
          <year>2019</year>
          .
          <article-title>BERT: Pre-training of deep bidirectional transformers for language understanding</article-title>
          .
          <source>In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</source>
          , Volume
          <volume>1</volume>
          (Long and Short Papers), pages
          <fpage>4171</fpage>
          -
          <lpage>4186</lpage>
          , Minneapolis, Minnesota, June. Association for Computational Linguistics.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          <string-name>
            <surname>M Ali</surname>
          </string-name>
          <article-title>Fauzi</article-title>
          and
          <string-name>
            <given-names>Anny</given-names>
            <surname>Yuniarti</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>Ensemble method for indonesian twitter hate speech detection</article-title>
          .
          <source>Indonesian Journal of Electrical Engineering and Computer Science</source>
          ,
          <volume>11</volume>
          (
          <issue>1</issue>
          ):
          <fpage>294</fpage>
          -
          <lpage>299</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          <string-name>
            <given-names>Ganesh</given-names>
            <surname>Jawahar</surname>
          </string-name>
          , Benoˆıt Sagot, and Djame´ Seddah.
          <year>2019</year>
          .
          <article-title>What does BERT learn about the structure of language? In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics</article-title>
          , pages
          <fpage>3651</fpage>
          -
          <lpage>3657</lpage>
          , Florence, Italy, July. Association for Computational Linguistics.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          <string-name>
            <surname>Sean</surname>
            <given-names>MacAvaney</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hao-Ren</surname>
            <given-names>Yao</given-names>
          </string-name>
          , Eugene Yang, Katina Russell, Nazli Goharian, and
          <string-name>
            <given-names>Ophir</given-names>
            <surname>Frieder</surname>
          </string-name>
          . 2019a.
          <article-title>Hate speech detection: Challenges and solutions</article-title>
          .
          <source>PloS one</source>
          ,
          <volume>14</volume>
          (
          <issue>8</issue>
          ):
          <fpage>e0221152</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          <string-name>
            <surname>Sean</surname>
            <given-names>MacAvaney</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hao-Ren</surname>
            <given-names>Yao</given-names>
          </string-name>
          , Eugene Yang, Katina Russell, Nazli Goharian, and
          <string-name>
            <given-names>Ophir</given-names>
            <surname>Frieder</surname>
          </string-name>
          . 2019b.
          <article-title>Hate speech detection: Challenges and solutions</article-title>
          .
          <source>PloS one</source>
          ,
          <volume>14</volume>
          (
          <issue>8</issue>
          ):
          <fpage>e0221152</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          <string-name>
            <given-names>Kristian</given-names>
            <surname>Miok</surname>
          </string-name>
          , Blaz Skrlj, Daniela Zaharie, and
          <string-name>
            <surname>Marko</surname>
          </string-name>
          Robnik-Sikonja.
          <year>2020</year>
          .
          <article-title>To ban or not to ban: Bayesian attention networks for reliable hate speech detection</article-title>
          . arXiv preprint arXiv:
          <year>2007</year>
          .05304.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          <string-name>
            <given-names>Matthew</given-names>
            <surname>Peters</surname>
          </string-name>
          , Mark Neumann, Mohit Iyyer, Matt Gardner, Christopher Clark,
          <string-name>
            <given-names>Kenton</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>and Luke</given-names>
            <surname>Zettlemoyer</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>Deep contextualized word representations</article-title>
          .
          <source>In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</source>
          , Volume
          <volume>1</volume>
          (
          <issue>Long Papers)</issue>
          , pages
          <fpage>2227</fpage>
          -
          <lpage>2237</lpage>
          , New Orleans, Louisiana, June. Association for Computational Linguistics.
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          <string-name>
            <given-names>Marco</given-names>
            <surname>Polignano</surname>
          </string-name>
          , Pierpaolo Basile, Marco de Gemmis, and
          <string-name>
            <given-names>Giovanni</given-names>
            <surname>Semeraro</surname>
          </string-name>
          .
          <year>2019</year>
          .
          <article-title>Hate speech detection through alberto italian language understanding model</article-title>
          .
          <source>In NL4AI@ AI* IA.</source>
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          <string-name>
            <given-names>Alec</given-names>
            <surname>Radford</surname>
          </string-name>
          , Karthik Narasimhan, Tim Salimans, and
          <string-name>
            <given-names>Ilya</given-names>
            <surname>Sutskever</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>Improving language understanding by generative pre-training.</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          <string-name>
            <given-names>Manuela</given-names>
            <surname>Sanguinetti</surname>
          </string-name>
          , Gloria Comandini, Elisa Di Nuovo, Simona Frenda, Marco Stranisci, Cristina Bosco, Tommaso Caselli, Viviana Patti, and
          <string-name>
            <given-names>Irene</given-names>
            <surname>Russo</surname>
          </string-name>
          .
          <year>2020</year>
          .
          <article-title>HaSpeeDe 2@EVALITA2020: Overview of the EVALITA 2020 Hate Speech Detection Task</article-title>
          . In Valerio Basile, Danilo Croce, Maria Di Maro, and Lucia C. Passaro, editors,
          <source>Proceedings of Seventh Evaluation Campaign of Natural Language Processing and Speech Tools for Italian. Final Workshop (EVALITA</source>
          <year>2020</year>
          ),
          <article-title>Online</article-title>
          . CEUR.org.
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          <string-name>
            <given-names>Gudbjartur</given-names>
            <surname>Ingi</surname>
          </string-name>
          Sigurbergsson and
          <string-name>
            <given-names>Leon</given-names>
            <surname>Derczynski</surname>
          </string-name>
          .
          <year>2019</year>
          .
          <article-title>Offensive language and hate speech detection for danish</article-title>
          . arXiv preprint arXiv:
          <year>1908</year>
          .04531.
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          <string-name>
            <given-names>Zeerak</given-names>
            <surname>Waseem</surname>
          </string-name>
          and
          <string-name>
            <given-names>Dirk</given-names>
            <surname>Hovy</surname>
          </string-name>
          .
          <year>2016</year>
          .
          <article-title>Hateful symbols or hateful people? predictive features for hate speech detection on twitter</article-title>
          .
          <source>In Proceedings of the NAACL student research workshop</source>
          , pages
          <fpage>88</fpage>
          -
          <lpage>93</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          <string-name>
            <given-names>Ziqi</given-names>
            <surname>Zhang</surname>
          </string-name>
          , David Robinson,
          <string-name>
            <given-names>and Jonathan</given-names>
            <surname>Tepper</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>Detecting hate speech on twitter using a convolution-gru based deep neural network</article-title>
          .
          <source>In European semantic web conference</source>
          , pages
          <fpage>745</fpage>
          -
          <lpage>760</lpage>
          . Springer.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>