<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>EVALITA 2020: Overview of the 7th Evaluation Campaign of Natural Language Processing and Speech Tools for Italian</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Valerio Basile</string-name>
          <email>valerio.basile@unito.it</email>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Maria Di Maro</string-name>
          <email>maria.dimaro2@unina.it</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Danilo Croce</string-name>
          <email>croce@info.uniroma2.it</email>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Lucia C. Passaro</string-name>
          <email>lucia.passaro@fileli.unipi.it</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>University of Naples “Federico II”</institution>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>University of Pisa</institution>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>University of Rome “Tor Vergata”</institution>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>University of Turin</institution>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2020</year>
      </pub-date>
      <abstract>
        <p>The Evaluation Campaign of Natural Language Processing and Speech Tools for Italian (EVALITA) is the biennial initiative aimed at promoting the development of language and speech technologies for the Italian language. EVALITA is promoted by the Italian Association of Computational Linguistics (AILC)1 and it is endorsed by the Italian Association for Artificial Intelligence (AIxIA)2 and the Italian Association for Speech Sciences (AISV)3. EVALITA provides a shared framework where different systems and approaches can be scientifically evaluated and compared with each other with respect to a large variety of tasks, suggested and organized by the Italian research community. The proposed tasks represent scientific challenges where methods, resources, and systems can be tested against shared benchmarks representing linguistic open issues or real world applications, possibly in a multilingual and/or multi-modal perspective. The collected data sets provide big opportunities for scientists to explore old and new problems concerning NLP in Italian as well as to develop solutions and to discuss the NLP-related issues within the community. Some tasks are traditionally present in the evaluation campaign, while others are completely new. This paper introduces the tasks proposed at EVALITA 2020 and provides an overview to the participants and systems whose descriptions and obtained results are reported in these Proceedings4. The EVALITA 2020 edition, held online on December 17th due to the COVID-19 pandemic, counts 14 different tasks. In particular, the selected tasks are grouped in five research areas (tracks) according to their objective and characteristics, namely (i) Affect, Hate, and Stance, (ii) Creativity and Style, (iii) New Challenges in Long-standing Tasks, (iv) Semantics and Multimodality, (v) Time and Diachrony. This edition was highly participated, with 51 groups whose participants have affiliation in 14 countries. Although EVALITA is generally promoted and targeted to the Italian research community, this edition saw an international participation, also thanks to the fact that several Italian researchers working in different countries contributed to the organization of the tasks or participated in them as authors. This overview is organized as follows: in Section 2 a brief description of the tasks belonging to the various areas is reported. Section 3 discusses the participation to the workshop referred to several aspects, from the research area, to the affiliation of authors. Section 4 describes the criteria used to assign the best system across tasks award, made by an ad-hoc committee starting from the suggestions of task organizers and reviewers. Finally, section 5 points out on both the obtained results and on the future of the workshop.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1 Introduction</title>
    </sec>
    <sec id="sec-2">
      <title>EVALITA 2020 Tracks and tasks</title>
      <p>
        In the 2020 edition of EVALITA, 14 different tasks were proposed, peer-reviewed and accepted. Data
were produced by the task organizers and made available to the participants. For the future availability
of this data we are going to release them on GitHub5, in accordance to the terms and conditions of the
respective data sources. Such a repository will also reference alternative repositories managed by the
task organizers. The tasks of EVALITA 2020 are grouped according to the following tracks:
AMI - Automatic Misogyny Identification
        <xref ref-type="bibr" rid="ref11">(Fersini et al., 2020)</xref>
        . This shared task is aimed at
automatically identifying misogynous content in Twitter for the Italian language. In particular, the AMI
challenge is focused on: (1) recognizing misogynous and aggressive messages and (2)
discriminating misogynistic contents from the non-misogynistic ones, while guaranteeing the fairness of the
model.
      </p>
      <p>
        ATE ABSITA - Aspect Term Extraction and Aspect-Based Sentiment Analysis
        <xref ref-type="bibr" rid="ref10 ref9">(De Mattei et al., 2020b)</xref>
        .
      </p>
      <p>A task on Aspect Term Extraction (ATE) and Aspect-Based Sentiment Analysis (ABSA). The task is
approached as a cascade of three subtasks: Aspect Term Extraction (ATE), Aspect-based Sentiment
Analysis (ABSA) and Sentiment Analysis (SA).</p>
      <p>
        HaSpeeDe - Hate Speech Detection
        <xref ref-type="bibr" rid="ref15">(Sanguinetti et al., 2020)</xref>
        . A rerun of the shared task on hate speech
detection at the message level on Italian social media texts proposed for the first time in 2018 for the
EVALITA evaluation campaign. The main task is a binary hate speech detection task, one in-domain
and one out-of-domain. On the same data provided for the main task, the topics of stereotypes in
communication and nominal utterances are investigated by of two pilot tasks.
      </p>
      <p>
        SardiStance - Stance Detection
        <xref ref-type="bibr" rid="ref7">(Cignarella et al., 2020)</xref>
        . The goal of the task is to detect the stance
of the author towards the target “Sardines movement” in Italian tweets. Two subtasks model (A)
Textual Stance Detection and (B) Contextual Stance Detection. Both the subtasks consist on a
threeclass (in favour, against, neutral) classification problem based on textual information only (A) or on
the text provided with additional information about the author and the post of the tweet.
      </p>
      <sec id="sec-2-1">
        <title>Creativity and Style</title>
        <p>
          CHANGE-IT - Style Transfer
          <xref ref-type="bibr" rid="ref10 ref9">(De Mattei et al., 2020a)</xref>
          . The first natural language generation task for
Italian. Change-IT focuses on style transfer performed on the headlines of two Italian newspapers
at opposite ends of the political spectrum. Specifically, the goal is to “translate” the headlines from
a style to another.
        </p>
        <p>
          TAG-it - Topic, Age and Gender Prediction
          <xref ref-type="bibr" rid="ref8">(Cimino et al., 2020)</xref>
          . TAG-IT is a profiling task for Italian.
        </p>
        <p>It is a follow-up of the GxG task organised in the context of EVALITA 2018. The task is aimed at
profiling along with three dimensions (Gender, Age, and Topic). Authors propose several subtasks
where participants are asked to predict one or more classes starting from the others.</p>
      </sec>
      <sec id="sec-2-2">
        <title>Semantics and Multimodality</title>
        <p>
          CONcreTEXT - Concreteness in Context
          <xref ref-type="bibr" rid="ref12">(Gregori et al., 2020)</xref>
          . The task focuses on automatic
assignment of concreteness values to words in context for the Italian and English language. Participants
are required to develop systems able to rate the concreteness of a target word in a sentence on a
scale from 1 (for fully abstract) to 5 (for maximally concrete).
        </p>
        <p>
          DANKMEMES - Multimodal Artefacts Recognition
          <xref ref-type="bibr" rid="ref14">(Miliani et al., 2020)</xref>
          . The first multimodal task
for Italian. The goal of the task is to deal with Italian memes considering textual and visual cues
together. Providing a corpus of memes on the 2019 Italian Government Crisis, DANKMEMES
features three subtasks: A) Meme Detection, B) Hate Speech Identification, and C) Event Clustering.
Ghigliottin-AI - Evaluating Artificial Players for the Language Game “La Ghigliottina”
          <xref ref-type="bibr" rid="ref2 ref3">(Basile et al.,
2020b)</xref>
          . The task challenges researchers to develop a system able to defeat human players at the
language game “La Ghigliottina”, which represents one of the most followed and appreciated quiz
games in Italy.
        </p>
        <p>
          PRELEARN - Prerequisite Relation Learning
          <xref ref-type="bibr" rid="ref1">(Alzetta et al., 2020)</xref>
          . The task is devoted to automatically
inferring prerequisite relations from educational texts. The task consists in classifying prerequisite
relations between pairs of concepts distinguishing between prerequisite pairs and non-prerequisite
pairs.
        </p>
      </sec>
      <sec id="sec-2-3">
        <title>Time and Diachrony</title>
        <p>
          DaDoEval - Dating Documents
          <xref ref-type="bibr" rid="ref13">(Menini et al., 2020)</xref>
          . The task focuses on assigning a temporal span to
a document, by recognising when a document was issued. A first one coarse-grained classification
subtask, participants are asked to provide a document with a class encoding the historical period (5
classes). The second Fine-grained classification task requires to attribute documents with a temporal
slice of 5 years.
        </p>
        <p>
          DIACR-Ita - Diachronic Lexical Semantics
          <xref ref-type="bibr" rid="ref2 ref3">(Basile et al., 2020a)</xref>
          . The first task on automatic
detection of lexical and semantic shift for Italian. The task challenges participants to develop systems
that can automatically detect if a given word has changed its meaning over time, given contextual
information from corpora.
        </p>
      </sec>
      <sec id="sec-2-4">
        <title>New Challenges in Long-standing Tasks</title>
        <p>
          AcCompl-it- Acceptability &amp; Complexity evaluation
          <xref ref-type="bibr" rid="ref5">(Brunato et al., 2020)</xref>
          . The task is aimed at
developing and evaluating methods to classify Italian sentences according to Acceptability and
Complexity. Given a set of sentences, two independent subtasks focus on predicting their acceptability
and complexity rate.
        </p>
        <p>
          KIPoS - Part-of-speech Tagging on Spoken Language
          <xref ref-type="bibr" rid="ref4 ref7">(Bosco et al., 2020)</xref>
          . The first task on Part of
Speech tagging of spoken language held at EVALITA. Benefiting from the KIParla corpus, a
resource of transcribed spoken Italian, the task provides three evaluation exercises focused on formal
versus informal spoken texts.
3
        </p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Participation</title>
      <p>
        EVALITA 2020 attracted the interest of a large number of researchers from academia and industry, for
a total of 51 teams composed of about 130 individuals participating in one or more of the 14 proposed
tasks. After the evaluation period, 58 system descriptions were submitted (reported in these proceedings),
i.e., a 70% percentage increase with respect to the previous EVALITA edition
        <xref ref-type="bibr" rid="ref6">(Caselli et al., 2018)</xref>
        .
      </p>
      <p>Moreover, task organizers allowed participants to submit more than one system result (called runs),
for a total of 240 submitted runs. Table 1 shows the different tracks and tasks along with the number of
participating teams and submitted runs. The data reported in the table is based on information provided
by the task organizers at the end of the evaluation process. Such data represents an overestimation with
respect to the systems described in the proceedings. The trends are similar, but there are differences due
to groups participating in more than a task, and groups that have not produced a system report.</p>
      <p>Differently from the previous EVALITA editions, the organizers were discouraged from distinguishing
the submissions between unconstrained and constrained runs6. The rationale for this decision is that the
recent spread and extensive use of pre-trained word embedding representations, especially as a strategy
to initialize Neural Network architectures, challenges this distinction at its very heart. Participation was
quite imbalanced across different tracks and tasks, as reported in Figure 1: each rectangle represents a
task whose size reflects the number of participants, while the color indicated the corresponding track.</p>
      <p>6A system is considered constrained when using the provided training data only; on the contrary, it is considered
unconstrained when using additional material to augment the training dataset or to acquire additional resources.
Creativity and Style</p>
      <p>In line with the previous editions of EVALITA, the track “Affect, Creativity and Style” covers about
half of the total in terms of participating teams. On the one hand, this demonstrates the well-known
interest of the NLP community for Social Media platforms and user-generated content. On the other hand,
we report a better balance with respect to the 2018 edition, where about 80% of the teams participated in
similar tracks (“Affect, Creativity and style” and “Hate Speech”, which have been merged in this edition).
Another significant number of teams participated to the “Semantics and Multimodality” and “Time and
Diachrony” tracks, while the other tracks where less participated. Unfortunately, no team participated to
the CHANGE-IT task, mainly due to the complexity of the task.</p>
      <p>In addition to being widely participated, the over 180 proceedings authors, including both participants
and task organizers, have affiliation in 18 countries, with the 64% from Italy and the 36% of participants
from Institutions and companies abroad. The group of the 59 task organizers have affiliations in 6
countries (90% from Italy while 10% from Institutions and companies abroad). The gender distribution
is highly balanced, with 30 females and males.</p>
    </sec>
    <sec id="sec-4">
      <title>Award: Best System Across Tasks</title>
      <p>In line with the previous edition, we confirmed the award to the best system across-task. The award
was introduced with the aim of fostering student participation to the evaluation campaign and to the
workshop. EVALITA received sponsorship funding from from Amazon Science, Bnova s.r.l., CELI s.r.l.,
the European Language and Resources Association (ELRA) and Google Research.</p>
      <p>A committee of 5 members was asked to choose the best system across tasks. Four of the five members
come from academia while the last one is from industry. The composition of the committee is balanced
with respect to the level of seniority as well as for their academic background (computer science-oriented
vs. humanities-oriented). In order to select a short list of candidates, the task organizers were invited
to propose up to two candidate systems participating to their tasks (not necessarily top ranking). The
committee was provided with the list of candidate systems and the criteria for eligibility, based on:
• novelty with respect to the state of the art;
• originality, in terms of identification of new linguistic resources, identification of linguistically
motivated features, and implementation of a theoretical framework grounded in linguistics;
• critical insight, paving the way to future challenges (deep error analysis, discussion on the limits of
the proposed system, discussion of the inherent challenges of the task);
• technical soundness and methodological rigor.</p>
      <p>We collected 10 system nominations from the organizers of 7 tasks from across all tracks. The
candidate systems are authored by 20 authors, among whom 12 are students, either at the master’s or PhD
level. The award recipient(s) will be announced during the final EVALITA workshop, during the plenary
session, held online.
5</p>
    </sec>
    <sec id="sec-5">
      <title>Final Remarks</title>
      <p>A record number of 14 tasks were organized at EVALITA 2020: some of them were revivals of tasks
in the past editions (such as Hate Speech Detection or Part-of-Speech Tagging), while others were
completely new (such as the ones involving Meme Processing or Stance Detection), and were greeted with
great enthusiasm by the NLP community.</p>
      <p>In this edition, the topics of Affect and Semantics were confirmed as two of the most interesting and
thriving ones, both in the number of organized tasks and actual participants. In any case, almost all
tasks involved the analysis of written texts. In fact, although the KIPoS task considered transcriptions of
spoken Italian utterances, no speech related tasks was proposed.</p>
      <p>Anyways, tasks concerning new problems and modalities have been proposed, such as the analysis
of memes, and two tasks oriented to the problem of time and diachrony. Moreover, this edition saw
an increase in tasks related to creativity and style, despite the fact that one one such tasks, namely
CHANGE-it, had no participation, probably due to its complexity and the lack of specific resources for
the task in the Italian community. Another task that received a rather low number of submission due to its
complexity is GhigliottinAI. Despite being a rather simple word-correlation problem by itself, it required
complex modelling of language and semantics to beat the challenge. A very interesting innovation for this
task was the evaluation framework, based on APIs, via a Remote Evaluation Server (RES). In general,
the most participated tasks have been those by which the linguistic problem could be modelled as a direct
classification or regression task.</p>
      <p>The competition attracted a record number of participating teams from academia and industry, for a
total of 51 teams and more than 180 authors with affiliations in 18 countries. Hopefully, this means
that EVALITA is becoming more and more popular also with foreign contributors, and it is becoming
an international workshop. First of all, this success confirms the beneficial impact of the organization
of the evaluation period based on non-overlapping windows (adopted from EVALITA 2018) in order to
help those who want to participate in more than one task. Moreover, we speculate that the technological
advancements and ease of use of existing open-source libraries for machine learning and natural language
processing improved the accessibility to the tasks, even for master students. In fact, we noticed an
increase in the participation of students, that contributed with state-of-the-art solutions to the tasks. We
can argue that the spread of frameworks such as PyTorch and Keras, together with pre-trained,
offthe-shelf language models, lowered the set-up costs to deal with complex NLP tasks. In general, we
noticed that most of the best systems are based on neural approaches. Among them, BERT or similar
Transformer-based architectures achieved the best results: more specifically, at least in 11 out of 14 tasks
best results (in at least one sub-task) were obtained by neural architectures based on or combined with</p>
      <sec id="sec-5-1">
        <title>Transformers.</title>
        <p>We are confident that the positive trends observed in this edition, concerning the participation and the
proliferation of tasks, has not yet reached a plateau. It would be auspicabile, among other aspects, to see
more tasks involving challenging settings such as, for example, multi-modal or multi-lingual analysis
involving Italian, in future EVALITA 2020 editions. Several areas represent fertile ground to organize
future tasks, such as domain adaptation (which was considered in previous editions of EVALITA), or
fewshot learning to support the definition of robust systems in challenging low-resource settings. Finally, we
believe in the importance of defining more structured tasks involving real applications to challenge the</p>
      </sec>
      <sec id="sec-5-2">
        <title>Italian community, e.g., Question Answering or Dialogue Agents.</title>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>Acknowledgments</title>
      <p>We would like to thank our sponsors Amazon Science7, Bnova8, CELI9, European Language and
Resources Association (ELRA)10 and Google Research11 for their support to the virtual event and to the
best-system across task award.</p>
      <p>Moreover, we gratefully acknowledge the members of the AILC board for their trust and support,
our EVALITA advisor Nicole Novielli and all the chairs of the 2018 edition, who helped us during the
organization process of EVALITA 2020. In addition, we sincerely thank the Best System across Tasks
committee for providing their expertise and experience.</p>
      <p>Finally, we know that EVALITA 2020 would not have been possible without the tireless effort,
enthusiasm, and originality of the task organizers, the colleagues and friends involved in running them, and all
the participants who contributed to make the workshop a success.
7https://www.amazon.science/
8https://www.bnova.it
9https://www.celi.it/
10http://elra.info/en/
11https://research.google/</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <string-name>
            <given-names>Chiara</given-names>
            <surname>Alzetta</surname>
          </string-name>
          , Alessio Miaschi, Felice Dell'Orletta,
          <string-name>
            <given-names>Frosina</given-names>
            <surname>Koceva</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Ilaria</given-names>
            <surname>Torre</surname>
          </string-name>
          .
          <year>2020</year>
          .
          <article-title>PRELEARN@EVALITA2020: Overview of the prerequisite relation learning task for italian</article-title>
          .
          <source>In Valerio Basile</source>
          , Danilo Croce, Maria Di Maro, and Lucia C. Passaro, editors,
          <source>Proceedings of the 7th evaluation campaign of Natural Language Processing</source>
          and
          <article-title>Speech tools for Italian (EVALITA 2020), Online</article-title>
          . CEUR.org.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          <string-name>
            <given-names>Pierpaolo</given-names>
            <surname>Basile</surname>
          </string-name>
          , Annalina Caputo, Tommaso Caselli, Pierluigi Cassotti, and
          <string-name>
            <given-names>Rossella</given-names>
            <surname>Varvara</surname>
          </string-name>
          .
          <year>2020a</year>
          .
          <article-title>DIACRIta@EVALITA2020: Overview of the evalita2020 diachronic lexical semantics (diacr-ita) task</article-title>
          . In Valerio Basile, Danilo Croce, Maria Di Maro, and Lucia C. Passaro, editors,
          <source>Proceedings of the 7th evaluation campaign of Natural Language Processing</source>
          and
          <article-title>Speech tools for Italian (EVALITA 2020), Online</article-title>
          . CEUR.org.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          <string-name>
            <given-names>Pierpaolo</given-names>
            <surname>Basile</surname>
          </string-name>
          , Marco Lovetere, Johanna Monti, Antonioa Pascucci, Federico Sangati, and
          <string-name>
            <given-names>Lucia</given-names>
            <surname>Siciliani</surname>
          </string-name>
          . 2020b.
          <article-title>Ghigliottin-AI@EVALITA2020: Evaluating artificial players for the language game “la ghigliottina”</article-title>
          . In Valerio Basile, Danilo Croce, Maria Di Maro, and Lucia C. Passaro, editors,
          <source>Proceedings of the 7th evaluation campaign of Natural Language Processing</source>
          and
          <article-title>Speech tools for Italian (EVALITA 2020), Online</article-title>
          . CEUR.org.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          <string-name>
            <given-names>Cristina</given-names>
            <surname>Bosco</surname>
          </string-name>
          , Silvia Ballare`,
          <string-name>
            <surname>Massimo</surname>
            <given-names>Cerruti</given-names>
          </string-name>
          , Eugenio Goria, and
          <string-name>
            <given-names>Caterina</given-names>
            <surname>Mauri</surname>
          </string-name>
          .
          <year>2020</year>
          .
          <article-title>KIPoS@EVALITA2020:overview of the task on kiparla part of speech tagging</article-title>
          . In Valerio Basile, Danilo Croce, Maria Di Maro, and Lucia C. Passaro, editors,
          <source>Proceedings of the 7th evaluation campaign of Natural Language Processing</source>
          and
          <article-title>Speech tools for Italian (EVALITA 2020), Online</article-title>
          . CEUR.org.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          <string-name>
            <given-names>Dominique</given-names>
            <surname>Brunato</surname>
          </string-name>
          , Cristiano Chesi, Felice Dell'Orletta, Simonetta Montemagni, Giulia Venturi, and
          <string-name>
            <given-names>Roberto</given-names>
            <surname>Zamparelli</surname>
          </string-name>
          .
          <year>2020</year>
          .
          <article-title>AcCompl-it@EVALITA2020: Overview of the acceptability &amp; complexity evaluation task for italian</article-title>
          .
          <source>In Valerio Basile</source>
          , Danilo Croce, Maria Di Maro, and Lucia C. Passaro, editors,
          <source>Proceedings of the 7th evaluation campaign of Natural Language Processing</source>
          and
          <article-title>Speech tools for Italian (EVALITA 2020), Online</article-title>
          . CEUR.org.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          <string-name>
            <given-names>T.</given-names>
            <surname>Caselli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Novielli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Patti</surname>
          </string-name>
          , and
          <string-name>
            <given-names>P.</given-names>
            <surname>Rosso</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>EVALITA 2018: Overview of the 6th Evaluation Campaign of Natural Language Processing and Speech Tools for Italian</article-title>
          . In Tommaso Caselli, Nicole Novielli, Viviana Patti, and Paolo Rosso, editors,
          <source>Proceedings of Sixth Evaluation Campaign of Natural Language Processing and Speech Tools for Italian. Final Workshop (EVALITA</source>
          <year>2018</year>
          ) , Turin, Italy. CEUR.org.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          <string-name>
            <given-names>Alessandra</given-names>
            <surname>Teresa</surname>
          </string-name>
          <string-name>
            <surname>Cignarella</surname>
          </string-name>
          , Mirko Lai, Cristina Bosco, Viviana Patti, and
          <string-name>
            <given-names>Paolo</given-names>
            <surname>Rosso</surname>
          </string-name>
          .
          <year>2020</year>
          .
          <article-title>SardiStance@EVALITA2020: Overview of the task on stance detection in italian tweets</article-title>
          . In Valerio Basile, Danilo Croce, Maria Di Maro, and Lucia C. Passaro, editors,
          <source>Proceedings of the 7th evaluation campaign of Natural Language Processing</source>
          and
          <article-title>Speech tools for Italian (EVALITA 2020), Online</article-title>
          . CEUR.org.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          <string-name>
            <given-names>Andrea</given-names>
            <surname>Cimino</surname>
          </string-name>
          , Felice Dell'Orletta,
          <string-name>
            <given-names>and Malvina</given-names>
            <surname>Nissim</surname>
          </string-name>
          .
          <year>2020</year>
          .
          <article-title>TAG-it@EVALITA2020: Overview of the topic, age, and gender prediction task for italian</article-title>
          .
          <source>In Valerio Basile</source>
          , Danilo Croce, Maria Di Maro, and Lucia C. Passaro, editors,
          <source>Proceedings of the 7th evaluation campaign of Natural Language Processing</source>
          and
          <article-title>Speech tools for Italian (EVALITA 2020), Online</article-title>
          . CEUR.org.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          <string-name>
            <surname>Lorenzo De Mattei</surname>
            , Michele Cafagna, Felice Dell'Orletta,
            <given-names>Malvina</given-names>
          </string-name>
          <string-name>
            <surname>Nissim</surname>
            ,
            <given-names>and Albert</given-names>
          </string-name>
          <string-name>
            <surname>Gatt</surname>
          </string-name>
          .
          <year>2020a</year>
          .
          <article-title>CHANGEIT@EVALITA2020: Change headlines, adapt news, generate</article-title>
          . In Valerio Basile, Danilo Croce, Maria Di Maro, and Lucia C. Passaro, editors,
          <source>Proceedings of the 7th evaluation campaign of Natural Language Processing</source>
          and
          <article-title>Speech tools for Italian (EVALITA 2020), Online</article-title>
          . CEUR.org.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          <string-name>
            <surname>Lorenzo De Mattei</surname>
            , Graziella De Martino, Andrea Iovine, Alessio Miaschi, Marco Polignano, and
            <given-names>Giulia</given-names>
          </string-name>
          <string-name>
            <surname>Rambelli</surname>
          </string-name>
          .
          <year>2020b</year>
          . ATE ABSITA@
          <article-title>EVALITA2020: Overview of the aspect term extraction and aspect-based sentiment analysis task</article-title>
          .
          <source>In Valerio Basile</source>
          , Danilo Croce, Maria Di Maro, and Lucia C. Passaro, editors,
          <source>Proceedings of the 7th evaluation campaign of Natural Language Processing</source>
          and
          <article-title>Speech tools for Italian (EVALITA 2020), Online</article-title>
          . CEUR.org.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          <string-name>
            <given-names>Elisabetta</given-names>
            <surname>Fersini</surname>
          </string-name>
          , Debora Nozza, and
          <string-name>
            <given-names>Paolo</given-names>
            <surname>Rosso</surname>
          </string-name>
          .
          <year>2020</year>
          .
          <article-title>AMI@EVALITA2020: Automatic misogyny identification</article-title>
          . In Valerio Basile, Danilo Croce, Maria Di Maro, and Lucia C. Passaro, editors,
          <source>Proceedings of the 7th evaluation campaign of Natural Language Processing</source>
          and
          <article-title>Speech tools for Italian (EVALITA 2020), Online</article-title>
          . CEUR.org.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          <string-name>
            <given-names>Lorenzo</given-names>
            <surname>Gregori</surname>
          </string-name>
          , Maria Montefinese, Daniele P. Radicioni, Andrea Amelio Ravelli, and
          <string-name>
            <given-names>Rossella</given-names>
            <surname>Varvara</surname>
          </string-name>
          .
          <year>2020</year>
          .
          <article-title>CONCRETEXT@EVALITA2020: The concreteness in context task</article-title>
          .
          <source>In Valerio Basile</source>
          , Danilo Croce, Maria Di Maro, and Lucia C. Passaro, editors,
          <source>Proceedings of the 7th evaluation campaign of Natural Language Processing</source>
          and
          <article-title>Speech tools for Italian (EVALITA 2020), Online</article-title>
          . CEUR.org.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          <string-name>
            <given-names>Stefano</given-names>
            <surname>Menini</surname>
          </string-name>
          , Giovanni Moretti, Rachele Sprugnoli, and
          <string-name>
            <given-names>Sara</given-names>
            <surname>Tonelli</surname>
          </string-name>
          .
          <year>2020</year>
          .
          <article-title>DaDoEval@EVALITA2020: Same-genre and cross-genre dating of historical documents</article-title>
          . In Valerio Basile, Danilo Croce, Maria Di Maro, and Lucia C. Passaro, editors,
          <source>Proceedings of the 7th evaluation campaign of Natural Language Processing</source>
          and
          <article-title>Speech tools for Italian (EVALITA 2020), Online</article-title>
          . CEUR.org.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          <string-name>
            <given-names>Martina</given-names>
            <surname>Miliani</surname>
          </string-name>
          , Giulia Giorgi, Ilir Rama, Guido Anselmi, and
          <string-name>
            <given-names>Gianluca E.</given-names>
            <surname>Lebani</surname>
          </string-name>
          .
          <year>2020</year>
          .
          <article-title>DANKMEMES@EVALITA2020: The memeing of life: memes, multimodality and politics</article-title>
          . In Valerio Basile, Danilo Croce, Maria Di Maro, and Lucia C. Passaro, editors,
          <source>Proceedings of the 7th evaluation campaign of Natural Language Processing</source>
          and
          <article-title>Speech tools for Italian (EVALITA 2020), Online</article-title>
          . CEUR.org.
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          <string-name>
            <given-names>Manuela</given-names>
            <surname>Sanguinetti</surname>
          </string-name>
          , Gloria Comandini, Elisa Di Nuovo, Simona Frenda, Marco Stranisci, Cristina Bosco, Tommaso Caselli, Viviana Patti, and
          <string-name>
            <given-names>Irene</given-names>
            <surname>Russo</surname>
          </string-name>
          .
          <year>2020</year>
          .
          <article-title>HaSpeeDe 2@EVALITA2020: Overview of the evalita 2020 hate speech detection task</article-title>
          .
          <source>In Valerio Basile</source>
          , Danilo Croce, Maria Di Maro, and Lucia C. Passaro, editors,
          <source>Proceedings of the 7th evaluation campaign of Natural Language Processing</source>
          and
          <article-title>Speech tools for Italian (EVALITA 2020), Online</article-title>
          . CEUR.org.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>