<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <article-id pub-id-type="doi">10.1007/978-3-030-03840-3_29</article-id>
      <title-group>
        <article-title>Leveraging LLMs for Event Extraction in Italian Documents: a Roadmap for Future Research</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Federica Rollo</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Giovanni Bonisoli</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Laura Po</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>"Enzo Ferrari" Engineering Department, University of Modena and Reggio Emilia</institution>
          ,
          <addr-line>MO 41121</addr-line>
          <country country="IT">Italy</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2024</year>
      </pub-date>
      <volume>6</volume>
      <fpage>29</fpage>
      <lpage>30</lpage>
      <abstract>
        <p>Event extraction is a task of significant interest in the field of Natural Language Processing (NLP) and plays a vital role in various applications, such as information retrieval and document summarization. Large Language Models (LLMs) have demonstrated remarkable capabilities in natural language understanding and generation. In this paper, we present a roadmap for the application of LLMs for event extraction from Italian documents, aiming to address the gap in research and resources for event extraction in non-English languages. We first discuss the challenges of event extraction and the current state-of-the-art approaches based on LLMs. Next, we present potential Italian datasets suitable for adapting linguistic models to the domain of event extraction. Furthermore, we outline future research directions and potential areas for improvement in this evolving ifeld.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;event extraction</kwd>
        <kwd>Large Language Model</kwd>
        <kwd>Italian language</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <sec id="sec-1-1">
        <title>2.2. Challenges</title>
      </sec>
      <sec id="sec-1-2">
        <title>2.3. Large Language Models based approaches</title>
        <sec id="sec-1-2-1">
          <title>Due to the complexity of the natural language, event</title>
          <p>extraction poses several challenges that require sophisti- Several approaches have been proposed for event
extraccated techniques to address efectively. tion in recent surveys, from traditional methods which</p>
          <p>
            The first challenge consists of detecting multiple rely on the use of linguistic rules for pattern
identificaevents described in the same document and understand- tion within the text to more advanced solutions such as
ing which are the references to each event. Natural lan- machine learning and deep learning algorithms able to
guage often contains ambiguous expressions that can learn patterns after training on annotated data, and the
refer to multiple events or entities. This ambiguity, along use of pre-trained language models [
            <xref ref-type="bibr" rid="ref11">2, 3</xref>
            ]. LLMs based
apwith the use of coreference, further complicates the task proaches have emerged as a promising avenue for event
of accurately extracting event data from text since resolv- extraction in recent years. These models leverage the
ing ambiguity requires contextual understanding and power of machine learning and deep learning algorithms
disambiguation techniques. as they are pre-trained on vast amounts of text data and
          </p>
          <p>Identifying relevant elements for each event requires then fine-tuned for specific tasks. By encoding contextual
distinguishing between event triggers (words or phrases information and capturing semantic relationships within
that indicate the occurrence of an event) and background the text, LLMs seem to be promising in identifying and
information and noise. Another complexity is given by extracting events from various sources.
the variability in language usage, writing styles, syntactic We identified three main approaches based on the use
structures, and document length. Indeed, event extrac- of LLMs that could reach good performance in event
extion can be performed on short text like tweets, longer traction: sequence labeling models, extractive Question
documents such as news articles, and lengthy documents Answering (QA) models and instruction-tuned models.
such as investigative reports or government documents.</p>
          <p>All these factors require the use of techniques able of Sequence Labeling models In Sequence labeling
accommodating these variations to achieve accurate and each token in a sequence is assigned a label based on
reliable results across diverse text types and genres. its role or category within the context of the sequence.</p>
          <p>Two of the key aspects of events are the time and the Sequence labeling models can be used to identify those
space, i.e., when the event took place and where. The text spans reporting relevant information within a text.
recognition and standardization of temporal and spatial Therefore, it is widely employed for several classical NLP
expressions could be complex since temporal reference tasks like part-of-speech (POS) tagging, named entity
can be expressed in various formats (such as dates, times, recognition (NER), text chunking.
part of the day). In addition, a document describing an Sequence labeling models are suitable for the scenario
event can refer to the location of the event providing of event extraction, where they can identify and classify
information at diferent granularity, for example indicat- those parts of text reporting information about events.
ing the name of the city, specifying the address, and/or Indeed, some works in literature have already treated
describing the type of the place like an apartment, a shop, event extraction as a sequence labeling or NER problem,
or a park. During event extraction, the references to all [4, 5], also for Italian Language [6].
these locations should be identified.</p>
          <p>Extractive Question Answering The goal of extrac- next-word prediction objective of LLMs and the users’
tive QA models is to understand an input question in objective of following their instructions helpfully and
natural language and extract the answer as a span from safely. Instruction-tuning involves a fine-tuning of
Autoan input text. QA models can facilitate rapid and efi- Regressive LLMs with input-output pairs, where input
cient access to event-related information by automati- denotes the human instructions, and output denotes the
cally identifying text spans containing the desired an- desired output that follows the instruction. The results
swers to specific questions. For instance, the question of this process are the Instruction-Tuned LLMs, designed
“When did the event take place?” (Q1) can be formulated specifically to provide appropriate results based on
into retrieve the date of the event. struction inputs. This ability is also enhanced as a
cross</p>
          <p>The results of these models depend significantly on the task generalization, leading Instruction-Tuned LLMs to
quality of the input documents, as well as the structure better performances on novel tasks.
of the questions provided to the models. Prior knowledge Instruction-Tuned LLMs can be employed to solve a
about the kind of event described in the document allows wide range of NLP tasks through various techniques of
to formulate ad hoc questions. For instance, considering prompt engineering [13], i.e., the process of designing
the document in Figure 1, the question “When did the air task-specific instructions to guide model output.
Therecrash take place?” (Q2) should provide more accurate an- fore, the utilization of these models can also yield benefits
swers than Q1. In addition, questions should be enriched for event extraction.
by other details about the event after a partial process of Currently, there are several Instruction-Tuned LLMs
event extraction. For example, the question “When did capable of understanding and generating text. For those,
the Flight 345 crash?” (Q3) contain the reference to the Italian represents a minority percentage in the training
lfight number and should help the QA models to select data compared to more widely used languages on the
the correct context for the extraction of the date. web such as English. Among these, there are proprietary</p>
          <p>
            Within the QA models, distinctions arise between models like GPT-3.5 and GPT-4 from OpenAI, Gemini
Single-Span QA (SQA) and Multi-Span QA (MQA). While from Google, and open-source families of LLMs like
Misthe former identifies a single text segment for each ques- tral [
            <xref ref-type="bibr" rid="ref20">14</xref>
            ] and Mixtral [15] from Mistral AI and Llama
tion, the latter locates answers even when distributed [11] and Llama 2 [16] from Meta. From this last family,
across non-consecutive text segments, potentially located Llamantino [17] has been derived through a language
far apart within a document. Given the prevalence of adaptation process to the Italian Language.
such scenarios, especially in complex inquiries and
detailed documents, the limitations of SQA models are
evident. An example is the annotation of “causalities and 3. Italian datasets
losses” in Figure 1. The recent surge in MQA model
development [
            <xref ref-type="bibr" rid="ref26">7, 8, 9</xref>
            ] underscores a notable interest. Currently, there are few Italian datasets suitable for event
          </p>
          <p>
            In the current state-of-the-art, the only Italian dataset extraction. Some of them provide a comprehensive
anproperly designed for training QA models is SQuAD-it notation of event-related data, while in other cases, only
[
            <xref ref-type="bibr" rid="ref25">10</xref>
            ], derived from the automatic translation of the En- one type of information (e.g., the temporal references) is
glish SQuAD dataset, consisting of a list of pairs question- annotated.
answer. However, this dataset can be used only for SQA,
therefore it is unsuitable for complex tasks like event 3.1. EVENTI
extraction which requires the ability to retrieve multiple
spans for one question.
          </p>
          <p>The EVENTI1 corpus was built in 2014 for the
evaluation of Temporal Information Processing systems of the
EVENTI evaluation exercise [18] in the EVALITA
workInstruction-Tuned models Among LLMs, Auto- shop. The corpus consists of three datasets: the Main
Regressive models such as GPT [1] or Llama [11] series task training data (274 documents) and test data (92
docstand out. These models leverage advanced deep learn- uments) of contemporary news articles and the Pilot task
ing techniques to predict the subsequent word based on (10 documents) test data of historical news articles. The
an input text. This prediction process is repeated mul- annotation guidelines involve the use of four tags to
antiple times, with each predicted word being added to notate diferent elements within news texts: the EVENT
the original text. By training on vast amounts of text tag is used to annotate all the mentions of events
includdata, Auto-Regressive LLMs efectively capture complex ing verbs, nouns, prepositional phrases and adjectives;
patterns and structures in language, leading them to gen- the TIMEX3 tag is used for temporal expressions; the
erate full and coherent text which is contextually relevant SIGNAL tag identifies textual items which encode a
reto input text. lation either between EVENTs, or TIMEX3s or both; the</p>
          <p>The research in recent years has led to the development
of instruction tuning [12] to bridge the gap between the 1https://sites.google.com/site/eventievalita2014/data-tools</p>
        </sec>
        <sec id="sec-1-2-2">
          <title>TLINK tag is used for temporal dependencies between EVENTs and/or Temporal Expressions.</title>
        </sec>
      </sec>
      <sec id="sec-1-3">
        <title>3.2. NewsReader MEANTIME</title>
        <sec id="sec-1-3-1">
          <title>The NewsReader MEANTIME (Multilingual Event ANd</title>
          <p>TIME) is a multilingual semantically annotated corpus of
480 Wikinews articles in four languages: English, Italian,
Spanish, and Dutch [19]. The corpus was released in 2016
and derives from the NewsReader Project2 [20] which
aims at extracting information about what happened to
whom, when, and where, processing a large volume of
ifnancial and economic data. The corpus is enriched with
annotations that span multiple levels, including entities,
entity mentions, events, temporal information, semantic
roles, and intra-document and cross-document event and
entity coreference.</p>
        </sec>
      </sec>
      <sec id="sec-1-4">
        <title>3.3. De Gasperi</title>
        <p>The De Gasperi corpus [21] is a collection of historical
documents by Alcide De Gasperi, the first Prime Minister
of the Italian Republic. The corpus was released in 2019
and includes 2,762 documents published between 1901
and 1954, originally released in an oral or written form. In
addition to the raw text, a set of meta-data and additional
semiautomatically annotated information are available.
The corpus contains diferent kinds of documents, like
daily press written by De Gasperi when he worked as a
journalist for newspapers in Trentino, and speeches in
institutional venues when he was a Member of the Italian
Parliament. In each document, references to persons and
places are annotated.
3.4. DICE
[24, 25]. The news articles underwent automated NLP
processes to extract temporal references, entities, and
corresponding DBpedia resources. Duplicates are annotated
to identify news articles referring to the same crime event.
The theft-related news articles are annotated manually
following a sophisticated annotation schema to identify
stolen items (What), crime locations (Where), references
to authors and victims, and their sociodemographic
characteristics (Who). The annotation provided in the dataset
is multi-span since it involves identifying and linking
multiple text spans within the document.</p>
      </sec>
      <sec id="sec-1-5">
        <title>3.5. EventNet-ITA</title>
        <p>EventNet-ITA4 [26] is an Italian corpus for Frame Parsing
applied to events released in 2024. Semantic Frame
Parsing is a task which aims at identifying semantic frames
within textual data. A semantic frame [27] is a cognitive
structure that organizes and represents knowledge about
a concept or situation. It consists of a set of
interconnected elements such as roles, attributes, and relations,
which collectively define the meaning and typical
features of that concept or situation. Frames help humans
understand and interpret language by providing a mental
framework for comprehending and categorizing
information.</p>
        <p>EventNet-ITA is built upon the idea of enabling frame
parsing for event extraction. It is composed of 53,854
sentences manually annotated with 205 semantic frames
of events and covers diferent domains, like conflictual,
social, communication, legal, geopolitical, economic and
biographical events.</p>
      </sec>
    </sec>
    <sec id="sec-2">
      <title>4. Future directions</title>
      <p>DICE3 [22] is a collection of 10,395 Italian news articles Automated information extraction from documents
condescribing crime events that happened in the Modena tinues to captivate the scientific community due to its
province between 2011 and 2021. The news articles are manifold advantages, facilitating improved information
extracted from one of the most popular local newspapers, accessibility across various domains. By leveraging LLMs
“Gazzetta di Modena”, following the approach described and exploiting annotated datasets, researchers can
dein [23]. Thanks to an agreement between the University velop robust event extraction systems capable of
achievof Modena and Reggio Emilia and the Gazzetta di Mod- ing high accuracy and eficiency across a wide range of
ena, DICE was released online in 2023, free to redistribute text sources. As the field continues to advance, further
and transform without encountering legal copyright is- research into LLMs and their applications in event
exsues under an Attribution-NonCommercial-ShareAlike traction is expected to drive continued innovation and
4.0 International (CC BY-NC-SA 4.0). progress in this area.</p>
      <p>Along with the data related to the title, the text, and Future directions will focus on three key aspects:
the publication date of each news article that are crawled
from the newspaper’s webpage, several annotations are
available on the data. The crime event category (e.g.,
theft, robbery) is assigned to each news article using text
categorization approaches based on word embeddings
• Definition of an Italian benchmark : while
we have identified five Italian datasets suitable
for event extraction, further eforts are needed to
expand their annotation and support
comprehensive event extraction tasks. This entails defining
2http://www.newsreader-project.eu/
3https://github.com/federicarollo/Italian-Crime-News
4https://huggingface.co/datasets/mrovera/eventnet-ita</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <article-title>a standardized benchmark for evaluating event</article-title>
          [2]
          <string-name>
            <given-names>G.</given-names>
            <surname>Frisoni</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Moro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Carbonaro</surname>
          </string-name>
          , A survey on
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          <source>comparisons between diferent approaches and Access</source>
          <volume>9</volume>
          (
          <year>2021</year>
          )
          <fpage>160721</fpage>
          -
          <lpage>160757</lpage>
          . doi:
          <volume>10</volume>
          .1109/
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          <article-title>fostering the development of more accurate</article-title>
          and
          <source>ACCESS</source>
          .
          <year>2021</year>
          .
          <volume>3130956</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          <article-title>reliable event extraction models</article-title>
          . [3]
          <string-name>
            <given-names>W.</given-names>
            <surname>Xiang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <article-title>A survey of event extraction</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          <article-title>• Evaluation of LLMs on the benchmark: de- from text</article-title>
          ,
          <source>IEEE Access 7</source>
          (
          <year>2019</year>
          )
          <fpage>173111</fpage>
          -
          <lpage>173137</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          <article-title>spite the limited literature on Italian event extrac</article-title>
          - doi:10.1109/ACCESS.
          <year>2019</year>
          .
          <volume>2956831</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          <article-title>tion, our preliminary evaluation of three BERT-</article-title>
          [4]
          <string-name>
            <given-names>A.</given-names>
            <surname>Ramponi</surname>
          </string-name>
          ,
          <string-name>
            <surname>R. van der Goot</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Lombardo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Plank</surname>
          </string-name>
          ,
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          <source>promising results [22]. However</source>
          , challenges per- in: B.
          <string-name>
            <surname>Webber</surname>
            , T. Cohn,
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>He</surname>
          </string-name>
          , Y. Liu (Eds.), Pro-
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          <article-title>sist, particularly related to the size and quality ceedings of the 2020 Conference on Empirical</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          <article-title>and comparing various approaches outlined in line,</article-title>
          <year>2020</year>
          , pp.
          <fpage>5357</fpage>
          -
          <lpage>5367</lpage>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          <article-title>Section 2.3. The evaluation will include the recent emnlp-main</article-title>
          .
          <volume>431</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          <article-title>Minerva models that represent the first</article-title>
          family of [5]
          <string-name>
            <given-names>S.</given-names>
            <surname>Pongpaichet</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Sukosit</surname>
          </string-name>
          , C. Duangtanawat,
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          <article-title>consuming process, new strategies will be studied news articles</article-title>
          ,
          <source>IEEE Access 12</source>
          (
          <year>2024</year>
          )
          <fpage>22778</fpage>
          -
          <lpage>22802</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          <article-title>to automate the process of annotation</article-title>
          . Employ- doi:10.1109/ACCESS.
          <year>2024</year>
          .
          <volume>3363879</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          <article-title>ing LLMs for data augmentation (i</article-title>
          .e., to expand [6]
          <string-name>
            <given-names>N.</given-names>
            <surname>Viani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T. A.</given-names>
            <surname>Miller</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Dligach</surname>
          </string-name>
          , S. Bethard,
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          <article-title>the spans to extract from the text (like “May 14th” ports, in: A. ten</article-title>
          <string-name>
            <surname>Teije</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Popow</surname>
            ,
            <given-names>J. H.</given-names>
          </string-name>
          <string-name>
            <surname>Holmes</surname>
          </string-name>
          ,
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          <article-title>ate a document with that span with the expected Springer International Publishing</article-title>
          , Cham,
          <year>2017</year>
          , pp.
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          <article-title>role in the event described (like “create a docu- 198-202</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          <article-title>ment describing an event that occurred</article-title>
          on May [7]
          <string-name>
            <given-names>H.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Tomko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Vasardani</surname>
          </string-name>
          , T. Baldwin, Mul-
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          <source>14th”)</source>
          .
          <article-title>This methodology allows for obtaining a tispanqa: A dataset for multi-span question an-</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          <article-title>Furthermore, this approach ofers control over Ruíz (Eds</article-title>
          .),
          <source>Proceedings of the 2022 Conference</source>
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          <article-title>opment of balanced and unbiased datasets essen- guage Technologies</article-title>
          ,
          <source>NAACL</source>
          <year>2022</year>
          , Seattle, WA,
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          <article-title>tial for training accurate and equitable AI models</article-title>
          .
          <source>United States, July 10-15</source>
          ,
          <year>2022</year>
          , Association for
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          <string-name>
            <given-names>Computational</given-names>
            <surname>Linguistics</surname>
          </string-name>
          ,
          <year>2022</year>
          , pp.
          <fpage>1250</fpage>
          -
          <lpage>1260</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          <source>doi:10</source>
          .18653/V1/
          <year>2022</year>
          .NAACL-MAIN.
          <year>90</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>E.</given-names>
            <surname>Segal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Efrat</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Shoham</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Globerson</surname>
          </string-name>
          , J. Be-
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          <string-name>
            <surname>rant</surname>
          </string-name>
          ,
          <article-title>A simple and efective model for answering [</article-title>
          1]
          <string-name>
            <given-names>T.</given-names>
            <surname>Brown</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Mann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Ryder</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Subbiah</surname>
          </string-name>
          ,
          <string-name>
            <surname>J. D.</surname>
          </string-name>
          multi-span questions,
          <year>2020</year>
          , p.
          <fpage>3074</fpage>
          -
          <lpage>3080</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          <string-name>
            <surname>Kaplan</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Dhariwal</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Neelakantan</surname>
            , P. Shyam, [9]
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Zhu</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Ahuja</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Juan</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          <string-name>
            <surname>Wei</surname>
            ,
            <given-names>C. K.</given-names>
          </string-name>
          <string-name>
            <surname>Reddy</surname>
          </string-name>
          ,
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          <string-name>
            <given-names>E.</given-names>
            <surname>Sigler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Litwin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Gray</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Chess</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Clark</surname>
          </string-name>
          ,
          <source>Linguistics: EMNLP</source>
          <year>2020</year>
          ,
          <string-name>
            <given-names>Online</given-names>
            <surname>Event</surname>
          </string-name>
          ,
          <fpage>16</fpage>
          -
          <lpage>20</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          <string-name>
            <given-names>C.</given-names>
            <surname>Berner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>McCandlish</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Radford</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Sutskever</surname>
          </string-name>
          ,
          <year>November 2020</year>
          , volume EMNLP 2020 of Findings
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          in: H.
          <string-name>
            <surname>Larochelle</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Ranzato</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Hadsell</surname>
          </string-name>
          , M. Bal- tics,
          <year>2020</year>
          , pp.
          <fpage>3840</fpage>
          -
          <lpage>3849</lpage>
          . doi:
          <volume>10</volume>
          .18653/V1/
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          <string-name>
            <surname>can</surname>
          </string-name>
          , H. Lin (Eds.),
          <source>Advances in Neural Information FINDINGS-EMNLP.342.</source>
        </mixed-citation>
      </ref>
      <ref id="ref33">
        <mixed-citation>
          <source>Processing Systems</source>
          , volume
          <volume>33</volume>
          ,
          <string-name>
            <surname>Curran</surname>
            <given-names>Associates</given-names>
          </string-name>
          , [10]
          <string-name>
            <given-names>D.</given-names>
            <surname>Croce</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Zelenanska</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Basili</surname>
          </string-name>
          , Neural learning
        </mixed-citation>
      </ref>
      <ref id="ref34">
        <mixed-citation>
          <string-name>
            <surname>Inc.</surname>
          </string-name>
          ,
          <year>2020</year>
          , pp.
          <fpage>1877</fpage>
          -
          <lpage>1901</lpage>
          .
          <article-title>for question answering in italian</article-title>
          , Lecture Notes
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>