<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Hate Speech Detection through AlBERTo Italian Language Understanding Model</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Marco Polignano</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Pierpaolo Basile</string-name>
          <email>pierpaolo.basile@uniba.it</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Marco de Gemmis</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Giovanni Semeraro</string-name>
          <email>giovanni.semeraro@uniba.it</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>University of Bari, Dept. Computer Science</institution>
          ,
          <addr-line>E.Orabona 4, 70125, Bari</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>The task of identifying hate speech in social networks has recently attracted considerable interest in the community of natural language processing. This challenge has great importance for identifying cyberattacks on minors, bullying activities, misogyny, or other kinds of hate discriminations that can cause diseases. Identifying them quickly and accurately can, therefore, help to solve situations that are dangerous for the health of the attacked people. Numerous national and international initiatives have addressed this problem by providing many resources and solutions to the problem. In particular, we focus on the Hate Speech Detection evaluation campaign (HaSpeeDe) held at Evalita 2018. It proposes an evaluation campaign with the aim of developing strategies for identifying hate speeches on Twitter and Facebook written in the Italian language. The dataset released for the task has been used by the classi cation approach proposed in this work for demonstrating that it is possible to solve the task e ciently and accurately. Our solution is based on an Italian Language Understanding model trained with a BERT architecture and 200M of Italian Tweets (AlBERTo). We used AlBERTo for ne-tuning a classi cation model of hate speech, obtaining state of the art results considering the best systems presented at the HaSpeeDe workshop. In this regard, AlBERTo is here proposed as one of the most versatile resources to be used for the task of classi cation of Social Media Textual contents in the Italian Language. The claim is supported by the similar results obtained by AlBERTo in the task of sentiment analysis, and irony detection demonstrated in previous works. The resources need for ne-tuning AlBERTo in these classi cation tasks are available at: https://github.com/marcopoli/AlBERTo-it</p>
      </abstract>
      <kwd-group>
        <kwd>Language Understanding Model</kwd>
        <kwd>AlBERTo</kwd>
        <kwd>Hate Speech</kwd>
        <kwd>Classi cation</kwd>
        <kwd>Machine Learning</kwd>
        <kwd>Deep Learning</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction and Motivations</title>
      <p>
        Hate speeches are characterized by their wide di usion on the web and by the
anonymity of the author, which makes this type of problem risky and relevant for
the community. These messages can be against groups of people, such as those
concerning discriminations about religion, race, and disability, or to a speci c
person. In addition, hate messages are characterized by di erent facets that are
very di erent from each other and which give rise to a wide and varied problem.
The interpretation of a message as hate or not is subject to a strong cultural and
social in uence by making the same message hatefully for some subjects (e.g.
them from a speci c country) and not hatefully for others. Hate Speech (HS) is,
consequently, a multi-faceted problem with strong cultural and social
intersections. The lexicon used in that messages is di cult to be found in a standard
dictionary and with many lexical variations making approaches of classi cation
based only on dictionaries unsuccessful. Therefore, the automatic identi cation
of hate messages is often a complex and intrinsically multidisciplinary task,
including the research areas of natural language processing (NLP), psychology,
law, social sciences, and many more. The hate speech detection is a challenging
task that gains hight interest by private industries and public institutions to be
able to remove potentially illegal contents quickly from the Web and to reduce
the connected risk to remove legal content unjustly. This has made it interesting
for us to apply an innovative classi cation model based on a language
understanding model for the Italian language (AlBERTo [
        <xref ref-type="bibr" rid="ref29">29</xref>
        ].) to obtain promising
results for the task.
      </p>
      <p>
        The classi cation model of this work is part of a wider national project,
\Contro l'odio"1, that aims to monitor, classify and summarize in statistics the
hate messages in Italian identi ed via Twitter. \Contro l'odio" is a project for
countering and preventing racist discrimination and HS in Italy, in particular
focused against immigrants. On the one hand, the project follows and extends
the research outcomes emerged from the `Italian Hate Map project' [
        <xref ref-type="bibr" rid="ref23">23</xref>
        ], whose
goal was to identify the most-at-risk areas of the Italian country, that is to say,
the areas where the users more frequently publish hate speech, by exploiting
semantic analysis and opinion mining techniques. On the other hand, \Contro
l'odio" bene ts from the availability of annotated corpora for sentiment analysis,
hate speech detection and related phenomena such as aggressiveness and o
ensiveness, to be used for training and tuning the HS detection tools [
        <xref ref-type="bibr" rid="ref27 ref31">31,27</xref>
        ]. The
project brings together the competences and active participation of civil society
organizations Acmos2 and Vox3, and two academic research groups, respectively
from the University of Bari and Turin.
2
      </p>
    </sec>
    <sec id="sec-2">
      <title>Related Work</title>
      <p>
        The interest of the scienti c community in the task of identifying hate speech and
related phenomena such as misogyny, cyberbullying, and abusive language has
been growing since 2016. Events such as HatEval 2019 [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], AMI at IberEval 2018
[
        <xref ref-type="bibr" rid="ref15">15</xref>
        ], HaSpeeDe 2018 [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] and AIM 2018 [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ] at EVALITA 2018 have contributed
to the emergence of a strong community of reference, methods, resources and
tools to address this complex task. For what concerns Italian a few resources
      </p>
      <sec id="sec-2-1">
        <title>1 https://controlodio.it/ 2 http://acmos.net/ 3 http://www.voxdiritti.it/</title>
        <p>
          have been recently developed drawn from Twitter [
          <xref ref-type="bibr" rid="ref27 ref31">31,27</xref>
          ] and Facebook [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ],
where the annotation of hateful contents also extends the simple markup of HS.
A multilingual lexicon of hate words has also been developed [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ] called Hurtlex4.
It is divided into 17 categories such as homophobic slurs, ethnic slurs, genitalia,
cognitive and physical disabilities, animals, and more.
        </p>
        <p>
          A recent survey of state of the art approaches for hate speech detection is
provided by Schmidt et al. [
          <xref ref-type="bibr" rid="ref33">33</xref>
          ]. The most common systems of speech detection
are based on algorithms of text classi cation that use a representation of contents
based on "surface features" such as them available in a bag of words (BOW)
[
          <xref ref-type="bibr" rid="ref11 ref34 ref36 ref37">11,37,36,34</xref>
          ]. A solution based on BOW is e cient and accurate, especially when
n-grams have been extended with semantic aspects derived by the analysis of the
text. [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ] describe an increase of the classi cation performances when features
such as the number of URLs, punctuations and not English words are added to
the vectorial representation of the sentence. [
          <xref ref-type="bibr" rid="ref35">35</xref>
          ] proposed, instead, to add as a
feature the number of positive, negative, and neutral words found in the sentence.
This idea demonstrated that the polarity of sentences positively supports the
classi cation task. These approaches su er from the lack of generalization of
words contained into the bag of words, especially when it is created through a
limited training set. In particular, terms found in the test sentences are often
missing in the bag. More recent works have proposed word embeddings [
          <xref ref-type="bibr" rid="ref19">19</xref>
          ]
as a possible distributional representation able to overcome this problem. This
representation has the advantage to transform semantically similar words into
a similar numerical vector (e.g. Word2Vec). Word embeddings are consequently
used by classi cation strategies such as Support Vector Machine and recently by
deep learning approaches such as deep recurrent neural networks [
          <xref ref-type="bibr" rid="ref20">20</xref>
          ].
        </p>
        <p>
          Limits of such technologies as Word2Vec [
          <xref ref-type="bibr" rid="ref22">22</xref>
          ], Glove [
          <xref ref-type="bibr" rid="ref25">25</xref>
          ], and FastText [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] fall
into the lack of use of context of terms when such representation is built
(contextfree). This means that each term has only a single wordembedding representation
in the distribution space, and di erent concepts related to the same term are not
represented. New strategies such as ELMO [
          <xref ref-type="bibr" rid="ref26">26</xref>
          ], GPT/GPT-2 [
          <xref ref-type="bibr" rid="ref30">30</xref>
          ], and BERT
[
          <xref ref-type="bibr" rid="ref14">14</xref>
          ] overcome this limit by learning a language understanding model for a
contextual and task-independent representation of terms. In particular, these models
are trained to predict the totality or a span of the starting sentence. This allows
obtaining a model able to predict, from a speci c context (often both previous
and subsequent), the most probable word from its vocabulary. Recently, several
articles have demonstrated the e ectiveness of this technique in almost all NLP
tasks in the English language, and recently, some multilingual models have been
distributed. This entails signi cant limitations related to the type of language
learned (related to the document style) and the limit of vocabulary extracted.
These reasons have led us to create the equivalent of the BERT model for the
Italian language and speci cally on the language style used on social networks:
alBERTo [
          <xref ref-type="bibr" rid="ref29">29</xref>
          ].
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>4 http://hatespeech.di.unito.it/resources.html</title>
        <p>The classi er proposed in this work about HS is based on AlBERTo,
demonstrating that its ne-tuned version is suitable for the task and it obtains better
results than them presented ad HaSpeeDe 2018 evaluation campaign.
3</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>AlBERTo-HS classi cation model</title>
      <p>
        The aim of this work is to create a classi cation model able to accurately classify
HS contents written in the Italian Language on Social Network such as
Facebook and Twitter. The analysis of state of the art shown that the main strategies
for facing these challenges, on the English language, are currently based on a
pre-trained language understanding models. Them, even in their multilingual
version, are not suitable for an use with data completely in a single language
and with a writing style di erent from that of books and encyclopedic
descriptions. It is well known that the language used on social networks is di erent from
the formal one as consequence of the presence of mentions, uncommon terms,
links, and hashtags that are not present elsewhere. AlBERTo [
        <xref ref-type="bibr" rid="ref29">29</xref>
        ] wants to be
the rst Italian language understanding model to represent a style of writing of
social networks, Twitter in particular, written in Italian. The ne-tuned
classication model proposed in this work is based AlBERTo derived by the software
code distributed through GitHub by Devlin et al. [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] 5 under the concession
of Google. It has been suitably modi ed to be learned without consequences on
text spans containing typical social media characters including emojis.
      </p>
      <p>
        The core deep learning structure of BERT and AlBERTo is a 12x
Transformer Encoder, where for each input, a percentage of terms is hidden and then
predicted for optimizing network weights in back-propagation. This strategy of
learning is commonly named "masked learning". In AlBERTo we implement
only the "masked learning" strategy, excluding the one based on "next following
sentence". This is a crucial aspect to be aware of because, in the case of tweets,
we do not have cognition of a ow of tweets as it happens in a dialog. For this
reason, we are sure enough that our AlBERTo is not suitable for the task of
question answering where this property is essential to have been learned by the
model. On the contrary, it is good enough to be used in tasks of classi cation
and predictions. In order to tailor the tweet text to BERT's input structure,
it has been necessary to carry out pre-processing operations. More speci cally,
using Python as the programming language, two libraries were mainly adopted:
Ekphrasis [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] and SentencePiece 6 [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ]. Ekphrasis is a famous tool for performing
an NLP pipeline on text extracted from Twitter. It has been used for:
{ Normalizing URL, emails, mentions, percents, money, time, date, phone
numbers, numbers, emoticons;
{ Tag and unpack hashtags.
      </p>
      <p>The normalization phase consists in replacing the term with a xed one in the
style of &lt; [entity type] &gt;. The tagging phase consists of annotating hashtags</p>
      <sec id="sec-3-1">
        <title>5 https://github.com/google-research/bert/ 6 https://github.com/google/sentencepiece</title>
        <p>by two tags &lt; hashtag &gt; ::: &lt; =hashtag &gt; representing its beginning and end
in the sentence. Whenever possible, the hashtag has been unpacked into known
words. For making the text clean and easily readable by the network, it has been
returned to its lowercase form and all characters except emojis, !, ? and accented
characters have been deleted.</p>
        <p>
          SentencePiece is a segmentation algorithm used for learning in an
unsupervised and language independent way the best strategy for splitting text into
terms for language models. It can process till 50k sentences per seconds and to
generate an extensive vocabulary. It includes in it the most common terms in the
training set and the subwords which occur in the middle of words, annotating
them with '##' in order to be able to encode also slang, incomplete or
uncommon words. SentencePiece produced also a tokenizer used for generating a list of
tokens for each tweet lately processed by the BERT "create pretraining data.py"
module. The dataset used for the learning phase of AlBERTo is TWITA [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ] a
huge corpus of Tweets in the Italian language collected from February 2012 to
September 2015 from Twitter o cial streaming API. In our con guration, we
randomly selected 200.000.000 of Tweets removing re-tweets , and processing
them with the pipeline of pre-processing previously described. The AlBERTo
classi cation model is the basis for any single-label or multi-label classi cation
task. For the speci c task of content classi cation of Hate speech, we will carry
out a subsequent phase of ne-tuning and adaptation of the model to
domainspeci c data. This allows us to obtain a classi er that exploits the language
knowledge obtained during the learning phase on the generic data and the
speci c domain characteristics learned during the ne-tuning phase. The ne-tuning
phase is con gured as a new training of AlBERTo with a number of epochs
sufciently small not to over t the model on the new data provided (usually from 3
to 15 epochs). This process allows us to vary the weights of the last layers of the
model in order to predict correctly the content provided in the testing phase. We
named the ne-tuned version of AlBERTo for Hate Speech as AlBERTo-HS.
4
        </p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Evaluation</title>
      <p>
        In order to evaluate Alberto-HS with contents produced by real users on social
networks, written in the Italian language, we decided to use the data released for
the evaluation campaign HaSpeeDe [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] at EVALITA 2018. This choice was made
considering that most of the available state of the art datasets are in English or
focused only on data collected from a single social media site such as Facebook,
Twitter, and others. The HaSpeeDe evaluation campaign was carried out by
dividing the problem into four di erent tasks:
{ HaSpeeDe-FB: where the goal is to train the model and predict if the
contents are HS on data extracted from Facebook;
{ HaSpeeDe-TW: where the goal is to train the model and predict if the
contents are of HS on data extracted from Twitter;
{ Cross-HaSpeeDe FB: where the goal is to train the model on data
collected from Facebook and predict if the contents are of HS on data extracted
from Twitter;
{ Cross-HaSpeeDe TW: where the goal is to train the model on data
collected from Twitter and predict if the contents are of HS on data extracted
from Facebook;
It is interesting to note that in the rst two tasks, the model must be able to
classify data coming from the same information source as the training phase.
Unlike the two "Cross" tasks, the data to be classi ed are di erent from those
used for the test, making the task of the classi er more challenging due to the
di erences in writing styles of the two platforms. In fact, not only are twitter
data shorter, containing mentions, hashtags, and retweets, but overall, they are
also less HS than Facebook data (only 32% compared to 68% for Facebook).
4.1
      </p>
      <p>
        Dataset and Metrics
Facebook dataset is collected from public pages on Facebook about
newspapers, public gures, artists and groups on heterogeneous topics. More than
17,000 comments were collected from 99 posts and subsequently annotated by
5 bachelor students. The nal dataset released consists of 3000 training phrases
(1618 not HS, 1382 HS) and 1000 test phrases (323 not HS, 677 HS).
Twitter dataset is part of the Hate Speech Monitoring program, coordinated
by the Computer Science Department of the University of Turin with the aim
at detecting, analyzing and countering HS with an inter-disciplinary approach
[
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]. Data were collected using keywords related to the concepts of immigrants,
Muslims and Rome. Data are annotated partly by experts and partly by Figure
Eight contributors. Also for this dataset 3000 training tweets were released (2028
not HS and 972 HS) and 1000 test tweets (676 not HS and 324 HS).
      </p>
      <p>The evaluation metrics used in HaSpeeDe campaign are the Precision, Recall
and F1-measure classics. Since the two classes (HS and not HS) are unbalanced
within the datasets, the F1 metric has been calculated separately on the two
classes and then macro-averaged.
4.2</p>
      <p>AlBERTo-HS</p>
      <p>ne-tuning
We ne-tuned AlBERTo two di erent times, in order to obtain one classi er for
each di erent dataset available as a training set. In particular, we created one
classi er for the HaSpeeDe-FB and the Cross-HaSpeeDe FB tasks using
Facebook training data and one for the HaSpeeDe-TW and the Cross-HaSpeeDe TW
using the Twitter training set. The ne-tuning learning phase has been run for
15 epochs, using a learning rate of 2e-5 with 1000 steps per loops on batches of
512 examples. The ne-tuning process was last 4 minutes every time.</p>
      <p>
        Systems and baseline
HaSpeeDe has received strong participation from the scienti c community and
therefore a large number of solutions to the task have been proposed [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ].
      </p>
      <p>
        GRCP [
        <xref ref-type="bibr" rid="ref24">24</xref>
        ] The authors developed a Bi-LSTM Deep Neural Network with
an Attention-based mechanism that allows to estimate the importance of each
word; the weight vector is then used with another LSTM model to classify the
text.
      </p>
      <p>
        HanSEL [
        <xref ref-type="bibr" rid="ref28">28</xref>
        ] The system proposed is based on an ensemble of three
classi cation strategies (Support Vector Machine with RBF kernel, Random Forest
and Deep Multilayer Perceptron), mediated by a majority vote algorithm. The
social media text is represented as a concatenation of word2vec vectors and a
TF-IDF bag of words.
      </p>
      <p>
        InriaFBK [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ] The authors implemented three di erent classi er models:
RNN, n-gram based and linear SVC.
      </p>
      <p>
        ItaliaNLP [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] Participants used a newly-introduced model based on a
2layer BiLSTM which exploits multi-task learning with additional data from the
2016 SENTIPOLC task [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ].
      </p>
      <p>
        Perugia [
        <xref ref-type="bibr" rid="ref32">32</xref>
        ] The participants' system uses a document classi er based on a
SVM algorithm. The features used by the system are a combination of FastText
word embeddings and other 20 syntactical features extracted from the text.
      </p>
      <p>
        RuG [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] The authors proposed two di erent classi ers: a SVM based on
linear kernel and an ensemble system composed of an SVM and a CNN combined
by a logistic regression meta-classi er.
      </p>
      <p>sbMMMP The authors tested two di erent systems. The rst one is based
on an ensemble of CNNs, whose outputs are then used as features by a
metaclassi er for the nal prediction. The second system uses a combination of CNN
and a GRU.</p>
      <p>
        StopPropagHate [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ] The authors use a classi er based on RNN with a
binary cross-entropy as loss function. In their system, each input word is
represented by a 10000-dimensional vector which is a one-hot encoding vector.
      </p>
      <p>
        VulpeculaTeam [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] According to the description provided by participants, a
neural network with three hidden layers was used, with word embeddings trained
on a set of previously extracted Facebook comments.
      </p>
      <p>For all tasks, the baseline score has been computed as the performance of a
classi er based on the most frequent class.
4.4</p>
      <p>Discussion of results
The evaluation of the results obtained by the AlBERTo-HS classi er proposed
in this work was carried out using the o cial evaluation script released at the
end of the campaign 7. Consequently, all the results obtained are replicable and
comparable with those present in the nal ranking of HaSpeeDe.</p>
      <sec id="sec-4-1">
        <title>7 http://www.di.unito.it/ tutreeb/haspeede-evalita18/data.html</title>
        <p>NOT HS HS
Precision Recall F-score Precision Recall F-score Macro-Avg F-score
0.2441
0.8410
0.8288
0.8002
0.7841
0.7751
0.7738
0.7554
0.7428
0.7147
0.6532
0.2424</p>
        <p>From the previous tables of results, it is possible to observe how
AlBERToHS succeeds in obtaining a state of the art results for two tasks out of four. The
di erences with other systems proposed in the evaluation campaign are about its
simplicity to be applied. A simple ne-tuning phase of AlBERTo on domain data
allows us to obtain very encouraging results. It is therefore interesting to note
that the entire process of pre-processing and ne-tuning lasts a few minutes, and
it can be used for obtaining excellent results for a wide variety of classi cation
tasks. In particular, the model is able to adapt in an excellent way to annotated
data (with the risk of over tting) producing excellent results if used in the same
application domain of the tuning phase. This is the case with the results obtained
for the HaSpeeDe-FB and HaSpeeDe-TW tasks.</p>
        <p>Looking at the results obtained for the classi cation of data coming from
Facebook (Tab. 1), it is possible to observe how the classi er is able to
capture the characteristics of the social language through the ne-tuning phase. In
particular, it is able to move its learned weights from them obtained parsing
the original training language based on Twitter to the one used on Facebook.
AlBERTo-HS obtains better performances than those of other participants in
the evaluation campaign, with regard to the precision in identifying the posts
not hate (0.8603), and the recall of those of hate (0.9453). The high value of
recall for hate messages allows us to assume that, on Facebook, they are
characterized by speci c thematics that make the classi cation task more inclusive
at the cost of accuracy, especially when not explicit hate messages are faced. As
an example, the message "Comunque caro Matteo se non si prendono
provvedimenti siamo rovinati." is classi ed as a hate message even if the annotators have
considered it to be not a hate message. In this example, it is clear that a basis
of hate is present in the ideas of the writer, even if it is not complicated by
what he writes. In other cases, words like "severe" have tricked the model into
classifying clearly neutral messages like the following as hate messages: "Matteo
sei la nostra voce!!! Noi donne non possiamo fare un cavolo! !! Leggi piu severe!".
Nevertheless, the average F1 score higher than 0.8410, show us that, unlike in
Twitter, the use of more characters available for writing allows people to be more
verbose and, therefore, more comfortable to identify. Table 2 shows the results
obtained for the classi cation of tweets. Here the values are not so di erent from
the rst in the ranking during the evaluation campaign Haspeede even if the
average value of F1 obtained of 0.8023 proves to be the best. This suggests that
the presence in the tweets of particular characters and implicitly of hate, the
brevity of the latter, and the increase in the number of ironic tweets make the
task more complicated than the previous one.</p>
        <p>As far as "Cross" classi cation problems are concerned, the results are not
guaranteed. In Tab. 3 it can be observed that the model has not been able to
correctly abstract from the domain data, obtaining not very good results for the
classi cation in a di erent domain. In particular, the model trained on Facebook
is able to obtain a score of 0.4750 of F1 on Twitter test data. A similar situation
is repeated for the results in Tab. 4 where for the task Cross-HaSpeeDe TW the
model is able to generalize slightly better than before but still gets the second
place in the ranking. These results con rm the di culty of the Cross tasks and
the drop in performance that is obtained through a transfer-learning strategy
like the one adopted here. The great di erences in writing styles used on the
two social networks do not allow the model to adapt properly to the domain
of application if ne-tuned on di erent stylistic data. So that AlBERTo is not
able to grasp those particularities of the language to be used in the classi cation
phase.</p>
        <p>In any case, we want to observe how it has been possible to obtain an
excellent result of classi cation by merely carrying out a phase of ne-tuning on the
model. To this end, we will consider as future works those of making a further
comparison with other language understanding models such as GPT2, XLNet,
RoBERTa trained on the Italian language with the aim of verifying if they can
be more robust to the changes in the writing style of the text to be classi ed.
5</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Conclusion</title>
      <p>The problem of hate speech is strongly perceived in online communities because
of its repercussions on the quality of life of hate victims. It is therefore of great
interest to both public and private organisations to be able to quickly identify
and remove hate messages. Numerous national and international initiatives have
been carried out in recent years, especially for the English language, leaving
the Italian language with few resources to address the problem. In this work
we have proposed a simple model of classi cation obtainable through a quick
ne-tuning phase of a wider language understanding model pre-trained on the
Italian language (AlBERTo). This model was evaluated on the data released
for the HaSpeeDe evaluation campaign held at the EVALITA 2018 workshop.
Data containing phrases extracted from Facebbok and Twitter were classi ed
according to four di erent tasks. The rst two involved training the model on
data from the same domain as the test data. On the contrary, the last two "Cross"
tasks involved a classi cation on data from a domain di erent from the training
one. The results obtained showed excellent performances when the model is
evaluated on data coming from the same distribution of training data. On the
contrary, good performances in this transfer learning task are not guaranteed
due to the great stylistic di erences of the language used on di erent online
platforms such as Facebook and Twitter. Future work will focus on the possibility
of learning a model that includes data from di erent online sources so as to make
it more complete and robust to stylistic variations.
6</p>
    </sec>
    <sec id="sec-6">
      <title>Acknowledgment</title>
      <p>This work is funded by project "DECiSION" codice raggruppamento: BQS5153,
under the Apulian INNONETWORK programme, Italy.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Bai</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Merenda</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zaghi</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Caselli</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nissim</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          : Rug@ evalita
          <year>2018</year>
          :
          <article-title>Hate speech detection in italian social media</article-title>
          . In: EVALITA@
          <string-name>
            <surname>CLiC-it</surname>
          </string-name>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Barbieri</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Basile</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Croce</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nissim</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Novielli</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Patti</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          :
          <article-title>Overview of the evalita 2016 sentiment polarity classi cation task</article-title>
          .
          <source>In: Proceedings of third Italian conference on computational linguistics</source>
          (CLiC-it
          <year>2016</year>
          )
          <article-title>&amp; fth evaluation campaign of natural language processing and speech tools for Italian</article-title>
          .
          <source>Final Workshop (EVALITA</source>
          <year>2016</year>
          )
          <article-title>(</article-title>
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Basile</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bosco</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fersini</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nozza</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Patti</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pardo</surname>
            ,
            <given-names>F.M.R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rosso</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sanguinetti</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Semeval-2019 task 5: Multilingual detection of hate speech against immigrants and women in Twitter</article-title>
          .
          <source>In: Proceedings of the 13th International Workshop on Semantic Evaluation</source>
          . pp.
          <volume>54</volume>
          {
          <fpage>63</fpage>
          . Association of Computational Linguistics (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Basile</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lai</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sanguinetti</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Long-term social media data collection at the university of turin</article-title>
          .
          <source>In: Fifth Italian Conference on Computational Linguistics</source>
          (CLiC-it
          <year>2018</year>
          ). pp.
          <volume>1</volume>
          {
          <issue>6</issue>
          .
          <string-name>
            <surname>CEUR-WS</surname>
          </string-name>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Bassignana</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Basile</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Patti</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          :
          <article-title>Hurtlex: A multilingual lexicon of words to hurt</article-title>
          .
          <source>In: Proceedings of the Fifth Italian Conference on Computational Linguistics</source>
          (CLiC-it
          <year>2018</year>
          ), Torino, Italy,
          <source>December 10-12</source>
          ,
          <year>2018</year>
          .
          <source>CEUR Workshop Proceedings</source>
          , vol.
          <volume>2253</volume>
          .
          <string-name>
            <surname>CEUR-WS.org</surname>
          </string-name>
          (
          <year>2018</year>
          ), http://ceur-ws.
          <source>org/</source>
          Vol-
          <volume>2253</volume>
          /paper49. pdf
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Baziotis</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pelekis</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Doulkeridis</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          : Datastories at semeval
          <article-title>-2017 task 4: Deep lstm with attention for message-level and topic-based sentiment analysis</article-title>
          .
          <source>In: Proceedings of the 11th International Workshop on Semantic Evaluation (SemEval2017)</source>
          . pp.
          <volume>747</volume>
          {
          <fpage>754</fpage>
          . Association for Computational Linguistics, Vancouver, Canada (
          <year>August 2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Bianchini</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ferri</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Giorni</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          :
          <article-title>Text analysis for hate speech detection in italian messages on twitter and facebook</article-title>
          .
          <source>In: Proceedings of the Sixth Evaluation Campaign of Natural Language Processing and Speech Tools for Italian. Final Workshop (EVALITA</source>
          <year>2018</year>
          )
          <article-title>co-located with the Fifth Italian Conference on Computational Linguistics (CLiC-it</article-title>
          <year>2018</year>
          ), Turin, Italy,
          <source>December 12-13</source>
          ,
          <year>2018</year>
          . (
          <year>2018</year>
          ), http://ceur-ws.
          <source>org/</source>
          Vol-
          <volume>2263</volume>
          /paper043.pdf
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Bojanowski</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Grave</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Joulin</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mikolov</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          :
          <article-title>Enriching word vectors with subword information</article-title>
          .
          <source>Transactions of the Association for Computational Linguistics</source>
          <volume>5</volume>
          ,
          <issue>135</issue>
          {
          <fpage>146</fpage>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Bosco</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Felice</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Poletto</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sanguinetti</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Maurizio</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          :
          <article-title>Overview of the evalita 2018 hate speech detection task</article-title>
          .
          <source>In: Proceedings of Sixth Evaluation Campaign of Natural Language Processing and Speech Tools for Italian. Final Workshop (EVALITA</source>
          <year>2018</year>
          ). vol.
          <volume>2263</volume>
          , pp.
          <volume>1</volume>
          {
          <issue>9</issue>
          .
          <string-name>
            <surname>CEUR</surname>
          </string-name>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Bosco</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Viviana</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bogetti</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Conoscenti</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          , Ru o, G.,
          <string-name>
            <surname>Schifanella</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Stranisci</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Tools and Resources for Detecting Hate and Prejudice Against Immigrants in Social Media</article-title>
          .
          <source>In: Proceedings of First Symposium on Social Interactions in Complex Intelligent Systems (SICIS)</source>
          ,
          <source>AISB Convention</source>
          <year>2017</year>
          ,
          <article-title>AI</article-title>
          and
          <string-name>
            <surname>Society</surname>
          </string-name>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhou</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhu</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Xu</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          :
          <article-title>Detecting o ensive language in social media to protect adolescent online safety</article-title>
          .
          <source>In: Privacy, Security, Risk and Trust (PASSAT)</source>
          ,
          <source>2012 International Conference on and 2012 International Confernece on Social Computing (SocialCom)</source>
          . pp.
          <volume>71</volume>
          {
          <fpage>80</fpage>
          .
          <string-name>
            <surname>IEEE</surname>
          </string-name>
          (
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Cimino</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>De Mattei</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dell'Orletta</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <article-title>: Multi-task learning in deep neural networks at evalita 2018</article-title>
          . In: EVALITA@
          <string-name>
            <surname>CLiC-it</surname>
          </string-name>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Del Vigna</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cimino</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dell'Orletta</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Petrocchi</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tesconi</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          : Hate Me, Hate Me Not:
          <article-title>Hate Speech Detection on Facebook</article-title>
          .
          <source>In: Proceedings of the First Italian Conference on Cybersecurity (ITASEC17)</source>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Devlin</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chang</surname>
            ,
            <given-names>M.W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lee</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Toutanova</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          : BERT:
          <article-title>Pre-training of deep bidirectional transformers for language understanding</article-title>
          .
          <source>In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</source>
          , Volume
          <volume>1</volume>
          (Long and Short Papers). pp.
          <volume>4171</volume>
          {
          <fpage>4186</fpage>
          . Association for Computational Linguistics, Minneapolis,
          <source>Minnesota (Jun</source>
          <year>2019</year>
          ), https://www.aclweb.org/anthology/N19-1423
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Fersini</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nozza</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rosso</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Overview of the evalita 2018 task on automatic misogyny identi cation (ami)</article-title>
          .
          <source>In: Proceedings of the 6th evaluation campaign of Natural Language Processing and Speech tools for Italian (EVALITA'18)</source>
          , Turin, Italy. CEUR. org (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <surname>Fersini</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nozza</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rosso</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Overview of the evalita 2018 task on automatic misogyny identi cation (AMI)</article-title>
          . In: Caselli,
          <string-name>
            <given-names>T.</given-names>
            ,
            <surname>Novielli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            ,
            <surname>Patti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            ,
            <surname>Rosso</surname>
          </string-name>
          , P. (eds.)
          <source>Proceedings of the Sixth Evaluation Campaign of Natural Language Processing and Speech Tools for Italian. Final Workshop (EVALITA</source>
          <year>2018</year>
          )
          <article-title>co-located with the Fifth Italian Conference on Computational Linguistics (CLiC-it</article-title>
          <year>2018</year>
          ), Turin, Italy,
          <source>December 12-13</source>
          ,
          <year>2018</year>
          .
          <source>CEUR Workshop Proceedings</source>
          , vol.
          <volume>2263</volume>
          . CEURWS.org (
          <year>2018</year>
          ), http://ceur-ws.
          <source>org/</source>
          Vol-
          <volume>2263</volume>
          /paper009.pdf
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          17.
          <string-name>
            <surname>Fortuna</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bonavita</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nunes</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Merging datasets for hate speech classi - cation in italian</article-title>
          .
          <source>In: Proceedings of the Sixth Evaluation Campaign of Natural Language Processing and Speech Tools for Italian. Final Workshop (EVALITA</source>
          <year>2018</year>
          )
          <article-title>co-located with the Fifth Italian Conference on Computational Linguistics (CLiC-it</article-title>
          <year>2018</year>
          ), Turin, Italy,
          <source>December 12-13</source>
          ,
          <year>2018</year>
          . (
          <year>2018</year>
          ), http://ceur-ws.
          <source>org/</source>
          Vol-
          <volume>2263</volume>
          /paper037.pdf
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          18.
          <string-name>
            <surname>Kudo</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          :
          <article-title>Subword regularization: Improving neural network translation models with multiple subword candidates</article-title>
          . arXiv preprint arXiv:
          <year>1804</year>
          .
          <volume>10959</volume>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          19.
          <string-name>
            <surname>Le</surname>
            ,
            <given-names>Q.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mikolov</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          :
          <article-title>Distributed representations of sentences and documents</article-title>
          .
          <source>In: International Conference on Machine Learning</source>
          . pp.
          <volume>1188</volume>
          {
          <issue>1196</issue>
          (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          20.
          <string-name>
            <surname>Mehdad</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tetreault</surname>
          </string-name>
          , J.:
          <article-title>Do characters abuse more than words?</article-title>
          <source>In: Proceedings of the 17th Annual Meeting of the Special Interest Group on Discourse and Dialogue</source>
          . pp.
          <volume>299</volume>
          {
          <issue>303</issue>
          (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          21.
          <string-name>
            <surname>Michele</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Stefano</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pinar</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sprugnoli</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Elena</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sara</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Serena</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          :
          <article-title>Comparing di erent supervised approaches to hate speech detection</article-title>
          .
          <source>In: EVALITA 2018</source>
          . pp.
          <volume>230</volume>
          {
          <fpage>234</fpage>
          . aAccademia University Press (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          22.
          <string-name>
            <surname>Mikolov</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sutskever</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Corrado</surname>
            ,
            <given-names>G.S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dean</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          :
          <article-title>Distributed representations of words and phrases and their compositionality</article-title>
          .
          <source>In: Advances in neural information processing systems</source>
          . pp.
          <volume>3111</volume>
          {
          <issue>3119</issue>
          (
          <year>2013</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          23.
          <string-name>
            <surname>Musto</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Semeraro</surname>
            , G., de Gemmis,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lops</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Modeling community behavior through semantic analysis of social data: The italian hate map experience</article-title>
          .
          <source>In: Proceedings of the 2016 Conference on User Modeling Adaptation and Personalization</source>
          . pp.
          <volume>307</volume>
          {
          <fpage>308</fpage>
          .
          <string-name>
            <surname>ACM</surname>
          </string-name>
          (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          24. De la Pen~a Sarracen,
          <string-name>
            <given-names>G.L.</given-names>
            ,
            <surname>Pons</surname>
          </string-name>
          ,
          <string-name>
            <surname>R.G.</surname>
          </string-name>
          ,
          <article-title>Mun~iz-</article-title>
          <string-name>
            <surname>Cuza</surname>
            ,
            <given-names>C.E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rosso</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Hate speech detection using attention-based lstm</article-title>
          . In: EVALITA@
          <string-name>
            <surname>CLiC-it</surname>
          </string-name>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          25.
          <string-name>
            <surname>Pennington</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Socher</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Manning</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          : Glove:
          <article-title>Global vectors for word representation</article-title>
          .
          <source>In: Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP)</source>
          . pp.
          <volume>1532</volume>
          {
          <issue>1543</issue>
          (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          26.
          <string-name>
            <surname>Peters</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Neumann</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Iyyer</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gardner</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Clark</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lee</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zettlemoyer</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          : Deep contextualized word representations pp.
          <volume>2227</volume>
          {
          <issue>2237</issue>
          (Jun
          <year>2018</year>
          ). https://doi.org/10.18653/v1/
          <fpage>N18</fpage>
          -1202, https://www.aclweb.org/ anthology/N18-1202
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          27.
          <string-name>
            <surname>Poletto</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Stranisci</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sanguinetti</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Patti</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bosco</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>Hate speech annotation: Analysis of an italian twitter corpus</article-title>
          .
          <source>In: Proceedings of the Fourth Italian Conference on Computational Linguistics</source>
          (CLiC-it
          <year>2017</year>
          ), Rome, Italy,
          <source>December 11-13</source>
          ,
          <year>2017</year>
          .
          <source>CEUR Workshop Proceedings</source>
          , vol.
          <year>2006</year>
          .
          <article-title>CEUR-WS.org (</article-title>
          <year>2017</year>
          ), http://ceur-ws.
          <source>org/</source>
          Vol-2006/paper024.pdf
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          28.
          <string-name>
            <surname>Polignano</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Basile</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          : Hansel:
          <article-title>Italian hate speech detection through ensemble learning and deep neural networks</article-title>
          . In: EVALITA@
          <string-name>
            <surname>CLiC-it</surname>
          </string-name>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          29.
          <string-name>
            <surname>Polignano</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Basile</surname>
            , P., de Gemmis,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Semeraro</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Basile</surname>
          </string-name>
          , V.:
          <article-title>AlBERTo: Italian BERT Language Understanding Model for NLP Challenging Tasks Based on Tweets</article-title>
          .
          <source>In: Proceedings of the Sixth Italian Conference on Computational Linguistics</source>
          (CLiC-it
          <year>2019</year>
          ).
          <source>CEUR</source>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          30.
          <string-name>
            <surname>Radford</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wu</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Child</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Luan</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Amodei</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sutskever</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          :
          <article-title>Language models are unsupervised multitask learners (</article-title>
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          31.
          <string-name>
            <surname>Sanguinetti</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Poletto</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bosco</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Patti</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Stranisci</surname>
            ,
            <given-names>M.:</given-names>
          </string-name>
          <article-title>An Italian Twitter Corpus of Hate Speech against Immigrants</article-title>
          .
          <source>In: Proceedings of the 11th Language Resources and Evaluation Conference</source>
          <year>2018</year>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          32.
          <string-name>
            <surname>Santucci</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Spina</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Milani</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Biondi</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          , Di Bari, G.:
          <article-title>Detecting hate speech for italian language in social media</article-title>
          .
          <source>In: EVALITA</source>
          <year>2018</year>
          ,
          <article-title>co-located with the Fifth Italian Conference on Computational Linguistics (CLiC-it</article-title>
          <year>2018</year>
          ). vol.
          <volume>2263</volume>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref33">
        <mixed-citation>
          33.
          <string-name>
            <surname>Schmidt</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wiegand</surname>
            ,
            <given-names>M.:</given-names>
          </string-name>
          <article-title>A survey on hate speech detection using natural language processing</article-title>
          .
          <source>In: Proceedings of the Fifth International Workshop on Natural Language Processing for Social Media</source>
          . pp.
          <volume>1</volume>
          {
          <issue>10</issue>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref34">
        <mixed-citation>
          34.
          <string-name>
            <surname>Sood</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Antin</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Churchill</surname>
          </string-name>
          , E.:
          <article-title>Profanity use in online communities</article-title>
          .
          <source>In: Proceedings of the SIGCHI Conference on Human Factors in Computing Systems</source>
          . pp.
          <volume>1481</volume>
          {
          <fpage>1490</fpage>
          .
          <string-name>
            <surname>ACM</surname>
          </string-name>
          (
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref35">
        <mixed-citation>
          35.
          <string-name>
            <surname>Van Hee</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lefever</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Verhoeven</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mennes</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Desmet</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>De Pauw</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Daelemans</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hoste</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          :
          <article-title>Detection and ne-grained classi cation of cyberbullying events</article-title>
          .
          <source>In: International Conference Recent Advances in Natural Language Processing (RANLP)</source>
          . pp.
          <volume>672</volume>
          {
          <issue>680</issue>
          (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref36">
        <mixed-citation>
          36.
          <string-name>
            <surname>Warner</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hirschberg</surname>
          </string-name>
          , J.:
          <article-title>Detecting hate speech on the world wide web</article-title>
          .
          <source>In: Proceedings of the Second Workshop on Language in Social Media</source>
          . pp.
          <volume>19</volume>
          {
          <fpage>26</fpage>
          .
          <article-title>Association for Computational Linguistics (</article-title>
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref37">
        <mixed-citation>
          37.
          <string-name>
            <surname>Xu</surname>
            ,
            <given-names>J.M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jun</surname>
            ,
            <given-names>K.S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhu</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bellmore</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Learning from bullying traces in social media</article-title>
          .
          <source>In: Proceedings of the 2012</source>
          conference
          <article-title>of the North American chapter of the association for computational linguistics: Human language technologies</article-title>
          . pp.
          <volume>656</volume>
          {
          <fpage>666</fpage>
          .
          <article-title>Association for Computational Linguistics (</article-title>
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>