=Paper= {{Paper |id=Vol-2765/99 |storemode=property |title=UNITOR @ Sardistance2020: Combining Transformer-based Architectures and Transfer Learning for Robust Stance Detection |pdfUrl=https://ceur-ws.org/Vol-2765/paper99.pdf |volume=Vol-2765 |authors=Simone Giorgioni,Marcello Politi,Samir Salman,Roberto Basili,Danilo Croce |dblpUrl=https://dblp.org/rec/conf/evalita/GiorgioniPS0C20 }} ==UNITOR @ Sardistance2020: Combining Transformer-based Architectures and Transfer Learning for Robust Stance Detection== https://ceur-ws.org/Vol-2765/paper99.pdf
        UNITOR @ Sardistance2020: Combining Transformer-based
       Architectures and Transfer Learning for Robust Stance Detection
    Simone Giorgioni, Marcello Politi, Samir Salman, Danilo Croce and Roberto Basili
          Department of Enterprise Engineering, University of Roma, Tor Vergata
                        Via del Politecnico 1, 00133 Roma, Italy
             {simone.giorgioni,marcello.politi,samir.salman}@alumni.uniroma2.eu
                                        {croce,basili}@info.uniroma2.it


                        Abstract                                1       Introduction
    English. This paper describes the UNI-                      Stance detection aims at detecting if the author of
    TOR system that participated to the Stance                  a text is in favor of a target topic, or against it (Kre-
    Detection in Italian tweets (Sardistance)                   jzl et al., 2017). In this task, a text pair is generally
    task within the context of EVALITA 2020.                    considered: one text expresses the topic, while the
    UNITOR implements a transformer-based                       other one reflects the author’s judgments. In a pos-
    architecture whose accuracy is improved                     sible variant to such a setting, the topic is implicit
    by adopting a Transfer Learning tech-                       within an entire document collection over which
    nique. In particular, this work investigates                the stance detection is applied.
    the possible contribution of three auxil-                      In this work, we will consider this last setting,
    iary tasks related to Stance Detection, i.e.,               as defined in the in the Stance Detection in Ital-
    Sentiment Detection, Hate Speech Detec-                     ian Tweets (SardiStance) task (Cignarella et al.,
    tion and Irony Detection. Moreover, UN-                     2020) within the EVALITA 2020 (Basile et al.,
    ITOR relies on an additional dataset auto-                  2020). A set of texts (here tweets) is provided,
    matically downloaded and labeled through                    almost all concerning the same topic, i.e., the Sar-
    distant supervision. The UNITOR system                      dines Movement1 . The goal is to recognize if each
    ranked first in Task A within the compe-                    tweet is for or against (or neither) such target, only
    tition. This confirms the effectiveness of                  exploiting textual information. According to the
    Transformer-based architectures and the                     task definition, this corresponds to the so-called
    beneficial impact of the adopted strategies.                Task A. This is quite challenging problem, since
                                                                it requires at the same time to discover if a text
    Italiano. Questo lavoro descrive UN-
                                                                refers to the target topic and the author’s orienta-
    ITOR, uno dei sistemi partecipanti
                                                                tion, only relying on short messages written in a
    allo Stance Detection in Italian tweet
                                                                very conversational style.
    (SardiStance) task. UNITOR implementa
                                                                   We thus present the UNITOR system partici-
    un’architettura neurale basata su Trans-
                                                                pating to the SardiStance task A. The system is
    former, la cui accuratezza viene miglio-
                                                                based on a Transformer-based architecture for text
    rata applicando un metodo di Transfer
                                                                classification (Devlin et al., 2019) that is directly
    Learning, che sfrutta le informazioni di tre
                                                                pre-trained over a large-scale document collection
    task ausiliari, ovvero Sentiment Detection,
                                                                written in Italian, namely UmBERTo. In a nut-
    Hate Speech Detection e Irony Detection.
                                                                shell, the adopted architecture, which has been
    Inoltre, l’addestramento di UNITOR puó
                                                                demonstrated achieving state-of-the-art results in
    contare su un insieme di dati scaricati ed
                                                                many NLP tasks (Devlin et al., 2019), takes in in-
    etichettati automaticamente applicando
                                                                put a message and associates it to one of the target
    un semplice metodo di Distant Supervi-
                                                                classes indicating the stance. Moreover, due to the
    sion. Il sistema si é classificato al primo
                                                                task complexity and the small size of the dataset,
    posto nella competizione, confermando
                                                                in order to improve the generalization capabili-
    l’efficacia delle architetture basate su
                                                                ties of the neural network, we adopted a Trans-
    Transformer e il contributo delle strategie
                                                                fer Learning approach (Pan and Yang, 2010). Our
    adottate.
                                                                main assumption is that Stance Detection is tied
     Copyright © 2020 for this paper by its authors. Use per-   to other tasks involving emotion and subjectivity
mitted under Creative Commons License Attribution 4.0 In-
                                                                    1
ternational (CC BY 4.0).                                                https://en.wikipedia.org/wiki/Sardines_movement
analysis (such as Sentiment Analysis or Irony De-        adopted auxiliary tasks are described in Section
tection) even though important differences do exist      2.2, while our Transfer leaning strategy is in Sec-
among them. As a simplified example, let us con-         tion 2.3. Finally, an automatic strategy for Data
sider a message such as “I like the Sardines Move-       Augmentation is presented in Section 2.4.
ment”: it clearly expresses a positive sentiment,
also being in favour of the target topic. However,       2.1    UNITOR as a Transformer-based
a message such as “I like the EVALITA campaign.”                Architecture
is positive as well but it does not express any sup-     The approach proposed in (Devlin et al., 2019),
port or opposition to the Sardines (and it should be     namely Bidirectional Encoder Representations
associated to the None class). We thus speculate         from Transformers (BERT) provides a very effec-
that an automatic system trained over an auxiliary       tive model to pre-train a deep and complex neu-
task (e.g., Sentiment Classification) is beneficial,     ral network over large scale collections of non an-
but the transfer process must be carefully designed      notated texts and to apply it to a large variety of
in order to avoid catastrophic forgetting or inter-      NLP tasks. The building block of BERT is the
ference problems (Mccloskey and Cohen, 1989).            Transformer element (Vaswani et al., 2017), an
   In this work, we investigate the possible contri-     attention-based mechanism that learns contextual
bution of three auxiliary tasks involving the recog-     relations between words in a text. BERT provides
nition of emotions according to different settings,      a sentence embedding (as well as the contextual-
i.e., Sentiment Detection and Classification, Hate       ized lexical embeddings of words in the sentence)
Speech Detection and Irony Detection. We adopt           through a pre-training stage aiming at the acquisi-
three different classifiers (one for each auxiliary      tion of an expressive and robust language and text
task) and use them to add additional information to      model. The Transformer reads the entire input se-
the tweets provided in the SardiStance dataset. As       quence of words at once and is optimized through
an example, when considering the auxiliary task          two pre-training tasks. The first pre-training ob-
involving Hate Detection, the corresponding clas-        jective is the (masked language modeling) (Devlin
sifier will augment each input tweet by expressing       et al., 2019). In addition, a Next Sentence Predic-
if this expresses hate or not. After this step, the      tion task is used to jointly pre-train text embed-
final classifier is expected to learn the association    dings able to soundly represent discourse level in-
between messages and the stance categories, “be-         formation. This last objective operates on text-pair
ing aware” (with some unavoidable noise) if the          representations and aims at modeling relational in-
message expresses some sort of hate, irony and           formation, e.g. between the consecutive sentences
more generally, sentiment. Finally, we investigate       in a text. On top of the produced embeddings,
the possibility of augmenting the training mate-         BERT applies a fine-tuning stage devoted to adapt
rial by automatically downloading messages and           the entire architecture to the targeted task.
labeling them through distant supervision (Go et            The fine-tuning process of BERT for sentence
al., 2009). We first selected few hashtags clearly       classification (here adopted) operates on a single
in favour (or not) of the target topic to download       texts or text pairs, which can be given in input to
and label a set of set of messages. Then, in order       BERT, in analogy with a next sentence prediction
to add a set of neutral messages, we selected a set      task. The special token [CLS] is used as first el-
of news titles concerning the Sardines Movement.         ement of each input sequence and the embedding
   The UNITOR system ranked first in the com-            produced by BERT are used in input to a linear
petition, suggesting that the combination of             classifier customized for the target classification
the Transformer-based learning with the adopted          task. While the BERT architecture is pre-trained
strategies of Transfer Learning and Data Augmen-         on large-scale corpora, its application to new tasks
tation is beneficial. In the rest of the paper, Sec. 2   is generally obtained by customizing the final clas-
describes UNITOR. In Sec. 3, the evaluations are         sifier to the targeted problem and fine-tuning all
reported while Sec. 4 derives the conclusions.           the network parameters for few epochs, to avoid
                                                         catastrophic forgetting. In (Liu et al., 2019b)
2   Transformer-based architectures and                  RoBERTa is proposed as a variant of BERT which
    Transfer Learning for Stance Detection               modifies some key hyperparameters, including re-
The UNITOR system implements a Transformer-              moving the next-sentence pre-training objective,
based architecture described in Section 2.1. The         and training on more data, with much larger mini-
batches and learning rates. This allows RoBERTa                Irony Detection. We speculate that a robust de-
to improve on the masked language modeling ob-                 tection of stance requires the recognition of irony,
jective compared with BERT and leads to better                 which can even reverse the output of the classi-
downstream task performances.                                  fication task. For example a false stance can be
   UNITOR is based on a RoBERTa architecture                   expressed through a ironic message, such as “Le
pre-trained over Italian texts: we adopted Um-                 Sardine sono il futuro passato dell’Italia”4 . The
BERTo2 which is pre-trained over a subset of the               objective of Irony Detection is to detect whether
OSCAR corpus, made of 11 billion tokens. These                 a given message is ironic or not. We used the
architectures achieved state-of-the-art results in a           dataset provided IronITA 2018 (Cignarella et al.,
wide range of NLP tasks. However, they also                    2018), where a dataset of 4, 800 labeled messages
rely on large scale annotated datasets composed                is made available. We adopted the original binary
of (possibly hundreds) thousands of examples. In               classification task, mapping ironic messages to the
order to improve the quality of this architecture in            and  labels.
the SardiStance Task with a quite limited dataset,             Hate Speech Detection. Being against a topic
we adopted a simple Transfer Learning strategy by              can be often expressed through messages express-
relying on the following three auxiliary tasks.                ing also hate. We thus introduce also the Hate
                                                               Speech Detection task, which involves the auto-
2.2    Supporting UNITOR through Auxiliary                     matic recognition of hateful contents. We con-
       tasks                                                   sidered the setting proposed in HaSpeeDe 2018
In this work, we speculate that the complexity of              (Bosco et al., 2018), where a dataset of 3, 000 mes-
the Stance detection task can be simplified when-              sages is made available. We adopted the original
ever the system to be trained is already aware if              binary classification task: we mapped messages
input messages express some sort of Sentiment,                 expressing hate with the  label and  in the other case.
information, we trained specific classifiers over
dedicated corpora made available in the previous               2.3    Transferring auxiliary tasks in the
editions of EVALITA, as it follows:                                   Transformer-based learning
Sentiment Detection and Classification. This                   In order to transfer the information from each aux-
task consists in the automatic detection of subjec-            iliary task into UNITOR, we first trained a spe-
tivity (and the eventual positive or negative polar-           cific UmBERTo-based sentence classifier on each
ity) in texts (Pang and Lee, 2008). Even though                of the datasets described in the previous section.
the Stance Detection is clearly different from a               In each case, the standard parameters proposed
traditional task of Sentiment Analysis, we spec-               in (Devlin et al., 2019) are used to fine-tune the
ulate that they are nevertheless related. As an                model5 . After these three training steps, the en-
example, we can suppose that the presence of                   tire SardiStance dataset is processed by each of the
stance is more probable in messages expressing                 three classifiers and the resulting labels are used to
subjectivity. We thus considered the setting pro-              “augment” the input messages. In particular, these
posed in SENTIPOLC 2016 (Barbieri et al., 2016)                labels generated a sort of new sentence, which is
where a dataset of 8, 000 tweets is made avail-                paired with the corresponding message. The fol-
able. For each message, the presence of subjec-                lowing example shows how a tweet6 against the
tivity is made explicit and, eventually, the posi-             movement is used in input to UNITOR:
tive and negative polarity. The labeling provided              “[CLS] negativo ironico odio [SEP]
in the dataset was slightly modified and mapped                #elezioniregionali Le Sardine aiuteranno a
to a classification problem over three classes: all            salvare il Paese! #mafammilpiacere Sono proprio
objective tweets were labeled with the special tag             dei bei perdigiorno falliti! [SEP]”
, the subjective and positive mes-
sages with  while the negative ones                  Consistently with (Devlin et al., 2019), the first
with 3 .                                                4
                                                                    In English: “Sardines are the future past of Italy”
                                                                  5
                                                                    The number of epochs was tuned over a development set
    2
      https://huggingface.co/Musixmatch/                       made of 10% of the corresponding dataset and the best epoch
umberto-commoncrawl-cased-v1                                   was selected by maximizing the classification accuracy.
    3                                                             6
      We discarded the few available messages with mixed po-        In English: “#regionalelections The Sardines will help to
larity, to simplify the final classification task.             save the country! #please They’re just a bunch of losers!”
pseudo-token [CLS] is added to generate the em-                   news. In the experimental evaluations discussed
bedding used in input in the final linear classi-                 in the next section, this dataset of “silver” data is
fier. Then, the pseudo-sentence “negativo iron-                   simply added to the training material. To avoid
ico odio” suggests that the message expresses neg-                over-fitting, we removed 90% of the occurrences
ative polarity and hate through the adoption of                   of the hashtags used as query in the new data.
irony. Finally, between the [SEP] pseudo-tokens,                  3       Results and Discussion
the original message is reported. This particular
                                                                  UNITOR participated to Task A - Textual Stance
schema resembles the classification of text pairs
                                                                  Detection (Cignarella et al., 2020) where the avail-
used in relational learning tasks, such as in Tex-
                                                                  able dataset is composed by 2,132 tweets con-
tual Entailment (Devlin et al., 2019). The output
                                                                  cerning the Sardines Movement: 1,028 tweets
of the auxiliary classifiers defines a sort of hypoth-
                                                                  are against the movement (label Against), 589
esis, i.e., the authors aims at expressing a negative
                                                                  tweets in favour of it (label Favour) and 515
sentiment through an ironic message which also
                                                                  tweets do not express any stance about the target
expresses hate, while the original message is the
                                                                  topic (label None).
direct consequence, i.e., the “implied” message7 .
                                                                     As discussed in Section 2, UNITOR is based
The UNITOR model is thus an UmBERTo-based
                                                                  on the UmBERTo pre-trained model, which re-
classifier trained over text pairs, where the first el-
                                                                  lies on the RoBERTa architecture. For parame-
ement encodes the information derived from the
                                                                  ter tuning, we adopted a 10-cross fold validation,
auxiliary tasks and the second one is the original
                                                                  so that the training material is divided in 10 folds,
message. Even though the quality of this label-
                                                                  each split according to 90%-10% proportion. The
ing process can introduce noise (due to incorrectly
                                                                  model is trained using a standard Cross-entropy
classified messages) this augmented input is ex-
                                                                  Loss and an ADAM optimizer initialized with a
pected to simplify the final training process, by
                                                                  learning rate set to 2 · 10−5 and linearly decreased
explicitly providing information about sentiment,
                                                                  during the training process. We trained the model
hate and irony.
                                                                  for 5 epochs, using a batch size of 32 elements.
2.4    Distant Supervision for Stance Detection                   At test time, an Ensemble of such classifiers is
                                                                  used: each message is in fact classified using all
In order to balance the limited amount of avail-
                                                                  10 models trained in the different folds and the la-
able data (especially considering the complexity
                                                                  bel suggested by the highest number of classifiers
of the task) we augmented the training material by
                                                                  is selected. In the Task A, we submitted two con-
labeling additional messages via Distant Supervi-
                                                                  strained runs, i.e., system considering only tweets
sion (Go et al., 2009). We speculate that a tweet
                                                                  from the competition, and two unconstrained ones,
containing an hashtag such as #vivalesardine (in
                                                                  where additional tweets were acquired and labeled
English: #ILikeSarine) is in favour to Sardines
                                                                  by applying the approach presented in Section 2.2.
instead of a tweet containing for example #sar-
                                                                  All models are implemented using Pytorch8 and
dinefritte (in English: #friedSardine) is against
                                                                  experiments were run over Google Colab9 .
to our target. Hence, we downloaded from the
                                                                     Results are reported in Table 1 in terms of Pre-
TWITA corpus (Basile and Nissim, 2013) 3, 200
                                                                  cision, Recall and F1 scores obtained by the dif-
tweets and labeled them via Distant Supervision.
                                                                  ferent models with respect to each label. The final
In particular, the following subset are derived:
                                                                  rank considers the average F1 (F1-avg) between
1, 500 tweets against the movement since contain-
                                                                  the Favour and Against classes.
ing #gatticonsalvini and 1,000 tweets in favour,
                                                                     First of all, the high complexity of this task is
since containing #nessunotocchilesardine, #ios-
                                                                  confirmed by the results obtained by the strong
toconlesardine, #unmaredisardine, #vivalesardine
                                                                  Baseline method (the last row). It is a Support
or #forzasardine. Finally, to enlarge the subset of
                                                                  Vector Machine trained over a simple Bag-of-
messages without stance, 700 neutral statements
                                                                  Word model (Cignarella et al., 2020) and achieves
were downloaded, which are actually titles from
                                                                  an average F1 of 57.84%, being competitive with
news, derived by querying “sardine” in Google
                                                                  many systems participating to the task and rank-
    7
      We investigate different ways to encode this information,   ing 13th over 22 submissions. One important re-
even using complex sentences, but negligible differences in
                                                                      8
the tuning process were measured, so we applied the simplest              https://pytorch.org/
                                                                      9
schema.                                                                   http://colab.research.google.com/
                                           F1                                 Rec                            Prec
 Rk     System
                           avg       Against Favor       None      Against   Favor      None      Against   Favor     None
   1    UNITOR_u_1        68.53%     78.66% 58.40%       39.10%    76.01%    57.65%    45.35%     81.50%    59.16%    34.36%
   2    UNITOR_c_1        68.01%     78.81% 57.21%       39.79%    74.66%    63.78%    43.60%     83.43%    51.87%    36.59%
   3    UNITOR_c_2        67.93%     79.39% 56.47%       36.72%    77.09%    61.22%    37.79%     81.83%    52.40%    35.71%
   4    Opponent_c_1      66.21%     75.80% 56.63%       42.13%    68.60%    64.29%    52.91%     84.69%    50.60%    35.00%
   5    UNITOR_u_2        66.06%     76.89% 55.22%       37.02%    72.64%    56.63%    44.77%     81.67%    53.88%    31.56%
   6    UmBERTo           65.69%     77.41% 53.97%       35.93%    74.12%    57.14%    40.11%     81.00%    51.14%    32.54%
  13    Baseline          57.84%     71.58% 44.09%       27.64%    68.06%    49.49%    29.65%     75.49%    39.75%    25.89%

Table 1: Results obtained by UNITOR at the SardiStance task. In bold best results for each measure. In
the system name "c" and "u" refer to constrained and unconstrained runs.

sult is obtained by the straight application of the               with respect to the best opponent system, which
UmBERTo model over the original messages (next                    achieved a 66.21% of F1. It seems that the noise
to last row in Table 1). In fact, this Transformer-               added both from the auxiliary tasks and the addi-
based architecture, empowered with the Ensem-                     tional data, negatively impacted the overall qual-
ble technique, achieves an average F1 of 65.69%:                  ity. On the contrary, when only the Hate Speech
a system which directly applies an Ensemble of                    Detection task is considered (i.e., UNITOR_u_1)
UmBERTo-based models would have ranked 6th                        additional data are positively capitalized by the
in the competition.                                               model, achieving the best average F1 score in the
   We thus trained UmBERTo by adopting the                        competition, i.e. 68.53%. These results suggest
Transfer Learning approach presented in Section                   that the combination of the Transformer-based
2.3 in the constrained setting. The adoption of                   learning with the adopted strategies of Transfer
all the three auxiliary tasks led to the constrained              Learning and Data Augmentation is highly ben-
submission called UNITOR_c_2. Moreover, we                        eficial, when only Hate is considered.
considered the training of UmBERTo by consid-                        From an error analysis, it seems that a signif-
ering one auxiliary task at a time. When consid-                  icant number of incorrect classifications occurred
ering only the Hate Speech Detection task, better                 in longer and complex messages, where the topic
results were obtained over the development set,                   of the stance is not clearly explicit nor captured
with respect to the adoption of the other tasks                   by the UmBERTo model, such as in “#carfagna:
taken individually, i.e., Sentiment Detection and                 “io per i liberali che non si affidano a Salvini” e
Irony Detection10 . Such a variant, called UN-                    “dalle sardine buone idee”. Auto-scacco in due
ITOR_c_1, considers tweets enriched only with                     mosse. Con la Polverini poi...”11 . This message is
information derived by the hate classifier and it                 considered to be Against while the system as-
generally shows higher precision with respect to                  signs the label None. Here, it is very challenging
the Against class. This suggests that a tweet                     to understand the connection between the “good
expressing hate is more likely in opposition to                   ideas of the sardines” and the very colloquial ex-
the Sardines Movement. Both constrained mod-                      pression “Auto-scacco” which can be translated as
els ranked 3rd and 2nd in the competition, respec-                “She messed herself ”. The same appears in the
tively. These results are impressive as they both                 tweet “Ho finalmente capito chi mi ricordava Mat-
outperformed of about 2% of absolute F1 the stan-                 tia Santori, quello delle sardine: Lodo Guenzi. (e
dard UmBERTo. Moreover, they confirm the ben-                     infatti in quanto a democristianitá stiamo lá)”12
eficial impact of Hate Speech Detection as an aux-                which again labeled Against but classified as
iliary task. Finally, we augmented the training                   None. Clearly the system is not able to link
dataset by using the additional data presented in                 the movement to its leader nor to the negative
Section 2.2. We extended the training material                    opinion about belonging to the Christian Demo-
used to train UNITOR_c_2 in order to obtain the                   crat Party. Another example is the tweet “Dopo
unconstrained submission called UNITOR_u_2. It                       11
                                                                        In English: “#carfagna: "come with me liberals who
is worth noticing that all three auxiliary tasks were             do not rely on Salvini" and "from Sardines movement good
used in this submission. This led to a performance                ideas." She messed herself up with two moves. Not to men-
drop, i.e. a 66.06% of average F1, which is lower                 tion Polverini...”
                                                                     12
                                                                        In English: “I finally understood who reminded me of
                                                                  Mattia Santori, the one with the Sardines movement: Lodo
   10
      The results of this tuning stage were not reported here     Guenzi. (in fact as far as Christian Democrats are concerned
for lack of space.                                                they are pretty the same).)”
avere ascoltato @luigidimaio mi viene in mente                Cristina Bosco, Felice Dell’Orletta, Fabio Poletto,
una sola parola:grazie. Fiducia nelle sue scelte                M. Sanguinetti, and M. Tesconi. 2018. Overview
                                                                of the evalita 2018 hate speech detection task. In
e immenso rispetto per i grandi risultati ottenuti.
                                                                EVALITA@CLiC-it.
Ora un nuovo inizio, con un nuovo entusiamo. An-
diamo versogli #statigenerali con serietà e matu-             Alessandra Teresa Cignarella, Simona Frenda, Valerio
                                                                Basile, Cristina Bosco, Viviana Patti, Paolo Rosso,
rità. Forza@mov5stelle!”13 . Here the system in-                et al. 2018. Overview of the EVALITA 2018 task
correctly assigns the Favour label because the                  on irony detection in Italian tweets (IronITA). In
tweet is in favour of a different movement.                     Sixth Evaluation Campaign of Natural Language
                                                                Processing and Speech Tools for Italian (EVALITA
4   Conclusion                                                  2018), volume 2263, pages 1–6.
In this work we present the results obtained by               Alessandra Teresa Cignarella, Mirko Lai, Cristina
the UNITOR system, which participated to the                    Bosco, Viviana Patti, and Paolo Rosso. 2020.
SardiStance task. UNITOR ranked first in Task                   SardiStance@EVALITA2020: Overview of the Task
                                                                on Stance Detection in Italian Tweets. In Valerio
A, both for constrained and unconstrained runs.                 Basile, Danilo Croce, Maria Di Maro, and Lucia C.
These results confirm the beneficial impact of                  Passaro, editors, Proceedings of the 7th Evalua-
Transformer based architecture for text classifi-               tion Campaign of Natural Language Processing and
cation also in the Stance Detection task. More-                 Speech Tools for Italian (EVALITA 2020). CEUR-
over, we demonstrate the beneficial impact of Hate              WS.org.
Speech Detection as an auxiliary task in a Transfer           Jacob Devlin, Ming-Wei Chang, Kenton Lee, and
Learning setting. Finally, we empirically demon-                 Kristina Toutanova. 2019. BERT: Pre-training of
                                                                 deep bidirectional transformers for language under-
strate that the adoption of Distance Supervision
                                                                 standing. In Proceedings of NAACL 2019, pages
is useful to reduce data sparseness. Future work                 4171–4186, Minneapolis, Minnesota, June.
will apply the above approaches to task B within
                                                              Alec Go, Richa Bhayani, and Lei Huang. 2009. Twit-
SardiStance. Moreover, we will investigate multi-               ter sentiment classification using distant supervision.
task learning approaches (Liu et al., 2019a) to cap-            Technical report.
italize information from auxiliary tasks in a more
                                                              Peter Krejzl, Barbora Hourová, and Josef Steinberger.
principled way.                                                 2017. Stance detection in online discussions.
                                                              Xiaodong Liu, Pengcheng He, Weizhu Chen, and Jian-
References                                                      feng Gao. 2019a. Multi-task deep neural networks
Francesco Barbieri, Valerio Basile, Danilo Croce,               for natural language understanding. In Proceedings
  Malvina Nissim, Nicole Novielli, and Viviana Patti.           of ACL, pages 4487–4496, Florence, Italy, July.
  2016. Overview of the evalita 2016 sentiment polar-         Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Man-
  ity classification task. In Proceedings of EVALITA            dar Joshi, Danqi Chen, Omer Levy, Mike Lewis,
  2016, Napoli, Italy, December 5-7, 2016, volume               Luke Zettlemoyer, and Veselin Stoyanov. 2019b.
  1749 of CEUR Workshop Proceedings.                            Roberta: A robustly optimized BERT pretraining
Valerio Basile and Malvina Nissim. 2013. Sentiment              approach. CoRR, abs/1907.11692.
  analysis on italian tweets. In Proceedings of the 4th       Michael Mccloskey and Neil J. Cohen. 1989. Catas-
  Workshop on Computational Approaches to Subjec-               trophic interference in connectionist networks: The
  tivity, Sentiment and Social Media Analysis, pages            sequential learning problem. The Psychology of
  100–107, Atlanta.                                             Learning and Motivation, 24:104–169.
Valerio Basile, Danilo Croce, Maria Di Maro, and Lu-          S.J. Pan and Q. Yang. 2010. A Survey on Transfer
  cia C. Passaro. 2020. Evalita 2020: Overview                   Learning. IEEE Transactions on Knowledge and
  of the 7th evaluation campaign of natural language             Data Engineering, 22(10):1345–1359.
  processing and speech tools for italian. In Valerio
  Basile, Danilo Croce, Maria Di Maro, and Lucia C.           Bo Pang and Lillian Lee. 2008. Opinion mining and
  Passaro, editors, Proceedings of Seventh Evalua-              sentiment analysis. Found. Trends Inf. Retr., 2(1-
  tion Campaign of Natural Language Processing and              2):1–135.
  Speech Tools for Italian. Final Workshop (EVALITA
  2020). CEUR-WS.org.                                         Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob
                                                                Uszkoreit, Llion Jones, Aidan N Gomez, Ł ukasz
   13
      In English: “After listening to @luigidimaio only one     Kaiser, and Illia Polosukhin. 2017. Attention is
expression came to my mind: thank you. I have trust in          all you need. In I. Guyon, U. V. Luxburg, S. Ben-
his choices and a huge respect for the great results ob-        gio, H. Wallach, R. Fergus, S. Vishwanathan, and
tained. Now it’s a new start, with new enthusiasm. Let’s        R. Garnett, editors, Advances in Neural Information
move towards the #statigenerali with seriousness and matu-      Processing Systems 30, pages 5998–6008.
rity.Forza@mov5stars”