<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Knowledge-based Contexts for Historical Named Entity Recognition &amp; Linking</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Emanuela Boros</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Carlos-Emiliano González-Gallardo</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Edward Giamphy</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ahmed Hamdi</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>José G. Moreno</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Antoine Doucet</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Preligens</institution>
          ,
          <addr-line>75009 Paris</addr-line>
          ,
          <country country="FR">France</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>University of La Rochelle</institution>
          ,
          <addr-line>L3i, 17000 La Rochelle</addr-line>
          ,
          <country country="FR">France</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>University of Toulouse</institution>
          ,
          <addr-line>IRIT, 31000 Toulouse</addr-line>
          ,
          <country country="FR">France</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>This paper summarizes the participation of the L3i laboratory of the University of La Rochelle in the Identifying Historical People, Places, and other Entities (HIPE) evaluation campaign of CLEF 2022 in both tasks: named entity recognition and classification (NERC), coarse- and fine-grained, and entity linking (EL) in historical newspapers and classical commentaries. For both tasks, we developed models based on our previous models, which ranked first at CLEF- hipe-2020. The NERC model is a Transformer-based architecture and the EL model is a BiLSTM-based architecture. For NERC, our main contribution is two-fold: (1) data-wise improvement - we propose a knowledge-based strategy to provide related context information to the NERC model; (2) model-wise improvement - we adapt the NERC model to the task of detecting coarse- and fine-grained entities in non-standard text via adapters and we include the knowledge-based contexts as context jokers. Our approaches ranked first on 84.6% of the leaderboards we participated in for NERC and 85.7% of them for EL.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;historical documents</kwd>
        <kwd>fine-grained named entity recognition</kwd>
        <kwd>named entity linking</kwd>
        <kwd>knowledge bases</kwd>
        <kwd>language models</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        The identification of entities in historical documents, such as people and places, can be seen
as a building block of historical knowledge that allows easier access and better information
retrieval [
        <xref ref-type="bibr" rid="ref1 ref2 ref3 ref4">1, 2, 3, 4</xref>
        ]. Also, knowledge about historical events is gradually fading, especially
among the younger generations. Thus, preserving the historical memory of the information
that can be extracted from historical documents and bringing them to a larger audience, not
limited to researchers and experts in the humanities [
        <xref ref-type="bibr" rid="ref5 ref6 ref7">5, 6, 7</xref>
        ], could lead to better and wider
access to cultural heritage.
      </p>
      <p>
        Although named entity recognition (NER) and linking (EL) systems have been developed
to process modern data collections in general, NER and EL systems for processing historical
documents are less common [
        <xref ref-type="bibr" rid="ref8 ref9">8, 9</xref>
        ]. Because these documents are not digitally born, they are
scanned and processed by optical character recognition (OCR) tools to extract their textual
content. However, the OCR process is not error free and misrecognizes some of the content.
This can be due to the level of degradation of the document being scanned, the digitization
process, and also the quality of the OCR tool. This causes digitization errors in the recognized
text, such as misspelled locations or names.
      </p>
      <p>
        In this context, the first CLEF-HIPE-2020 edition [
        <xref ref-type="bibr" rid="ref10 ref11 ref12">10, 11, 12</xref>
        ] proposed the tasks of named
entity recognition and classification (NERC), both fine- and coarse-grained, and entity linking
(EL) in historical newspapers written in English, French and German. The evaluation showed
that neural-based systems with pre-trained language models or Transformer-based approaches
[
        <xref ref-type="bibr" rid="ref13 ref14 ref15">13, 14, 15</xref>
        ] clearly prevailed in NERC [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ], beating symbolic conditional random field (CRF)
[
        <xref ref-type="bibr" rid="ref17 ref18">17, 18</xref>
        ], pattern-based approaches or BiLSTMs [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ] by a large margin.
      </p>
      <p>
        For its second edition, the HIPE evaluation campaign1 took advantage of the availability
of several NE annotated datasets produced by several European cultural heritage projects
[
        <xref ref-type="bibr" rid="ref20">20</xref>
        ]. In this paper, we present our participation in the Identifying Historical People, Places, and
other Entities (HIPE) evaluation campaign of CLEF 2022 in both tasks: NERC, fine-grained
and coarse-grained, and EL in historical newspapers. For both tasks, we based our models on
those that we proposed at CLEF-HIPE-2020 [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ]. The NERC model was mainly based on the
Transformer architecture [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ] and the EL model was based on a BiLSTM architecture [
        <xref ref-type="bibr" rid="ref22">22</xref>
        ]. For
NERC, our main contribution is two-fold: (1) we propose a knowledge-based system, where
we build a multilingual knowledge base resting on Wikipedia and Wikidata to provide related
context information to the NERC model (data-wise improvement); (2) we adapt the NERC
model to the task of detecting coarse- and fine-grained entities in non-standard text by learning
modular language- and task-specific representations via newly-proposed additional adapters
[
        <xref ref-type="bibr" rid="ref23 ref24">23, 24</xref>
        ], small bottleneck layers inserted between the weights of two auxiliary Transformer
layers (model-wise improvement) [
        <xref ref-type="bibr" rid="ref25">25</xref>
        ]. Furthermore, for taking advantage of the additional
Wikipedia-based contexts, we include them in the model with mean-pooled representations
that we refer to as context jokers. Oficial results of our participation show the efectiveness of
our models over the CLEF-HIPE-2022 benchmark.
      </p>
      <p>The paper is organized as follows: Section 2 introduces the task and the datasets. Section 3
presents our knowledge-retrieval modules. Sections 4 and 5 respectively present our NERC and
EL systems and their corresponding performance. Conclusions are drawn in Section 6, where
future work is also presented.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Datasets</title>
      <p>The CLEF-HIPE-2022 competition proposed corpora composed of historical newspapers and
classical commentaries covering circa 200 years. The historical newspaper data is composed of
ifve datasets in English, Finnish, French, German and Swedish which originate from various
projects and national libraries in Europe, from which, we experimented with the hipe-2020
dataset. hipe-2020 includes newspaper articles from Swiss, Luxembourgish and American</p>
      <sec id="sec-2-1">
        <title>1https://hipe-eval.github.io/HIPE-2022/</title>
        <p>
          newspapers in French, German, and English (19C-20C) and it contains 19,848 linked entities
[
          <xref ref-type="bibr" rid="ref10 ref11 ref12">10, 11, 12</xref>
          ].
        </p>
        <p>
          We also experimented with the classical commentaries data from the Ajax Multi-Commentary
(ajmc) project that is composed of digitized 19C commentaries published in French, German, and
English [
          <xref ref-type="bibr" rid="ref26">26</xref>
          ], annotated with both universal named entities (person, location, organisation) and
domain-specific named entities (bibliographic references to primary and secondary literature).
        </p>
        <p>Table 1 presents the statistics regarding the number and type of entities in the aforementioned
datasets divided according to the training, development, and test sets.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Knowledge-based Contexts</title>
      <p>
        One of the main challenges of NER applied to historical newspapers and classical commentaries
concerns the digitization process of these heritage materials. The OCR output contains errors
which produce noisy text and complications, similar to those studied in [
        <xref ref-type="bibr" rid="ref27">27</xref>
        ]. Introducing
external grammatically correct contexts into NERC systems have been shown to have a positive
impact over the entities identification [
        <xref ref-type="bibr" rid="ref28">28</xref>
        ]. It consists on adding complementary and related
sentences, paragraphs or documents from external resources like Wikipedia or knowledge
graphs (KG) to enrich the surrounding of an entity, which helps NERC systems on detecting
the correct label. KGs structure information in a connected form, by representing entities (e.g.,
people, places) as nodes, and relationships between entities (e.g., being part of, being located in)
as edges. Thus, we propose two main techniques for generating additional contexts:
• Wikipedia Knowledge Retrieval Module: We create a local instance of ElasticSearch2, which
provides dense vector field indexing and a -nearest neighbor (kNN) search API. Given
a query vector, this API obtains the  closest vectors and returns those documents as
search hits.
• Knowledge Graph Embedding Retrieval Module: We produce English contexts by extending
the indexing scheme to a knowledge graph embedding model over the Wikidata5m3 [
        <xref ref-type="bibr" rid="ref29">29</xref>
        ]
dataset.
      </p>
      <sec id="sec-3-1">
        <title>3.1. Wikipedia Knowledge Retrieval Module</title>
        <p>
          We download the latest (02/04/2022) XML dumps4 of the French and German Wikipedia and
transform them into plain text using the Wikipedia2Vec [
          <xref ref-type="bibr" rid="ref30">30</xref>
          ] utility5. We focus on French
and German since for English we create another type of retrieval module which also contains
Wikipedia paragraphs. Similar to Wang et al. [
          <xref ref-type="bibr" rid="ref28">28</xref>
          ], we define a document, inside our instance
of ElasticSearch, as a triplet composed of a sentence, a title, and a paragraph. We create a
dense vector index over the sentence embedding field computed with a pre-trained multilingual
Sentence-BERT model6 [
          <xref ref-type="bibr" rid="ref31 ref32">31, 32</xref>
          ]. During context retrieval, for a given sentence from the datasets
described in Section 2, we compute its dense vector representation with the same multilingual
Sentence-BERT pre-trained model and take it as a query to retrieve the top-k semantically
similar documents based on a k-nearest neighbors algorithm (k-NN) cosine similarity search
over the sentence embedding field (Figure 1).
        </p>
        <sec id="sec-3-1-1">
          <title>2We utilized ElasticSearch v8.1.</title>
          <p>3https://deepgraphlearning.github.io/project/wikidata5m
4https://dumps.wikimedia.org/
5https://wikipedia2vec.github.io/wikipedia2vec/
6https://huggingface.co/sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2</p>
        </sec>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Knowledge Graph Embedding Retrieval Module</title>
        <p>
          Wikidata5m is a large-scale KG with aligned entity descriptions. It integrates around five million
Wikidata7 entities, which are described in the first paragraph of the corresponding Wikipedia
pages. We index the Wikidata5m dataset along with the dense vectors produced by the RotatE
KG embedding model [
          <xref ref-type="bibr" rid="ref33">33</xref>
          ] pre-trained over the same dataset8. RotatE defines each relation
between entities as a rotation from the source entity to the target entity in the dense vector
space. In this case, we describe “an ElasticSearch document” as a triplet formed by an entity
identifier, an entity description, and an entity embedding. We create a standard index on the
entity identifier field and two dense vector indexes: the former on the entity embedding field,
and the latter on the embeddings from the entity description field obtained with the same
Sentence-BERT model as in the previous module. We propose two diferent methods for context
retrieval (Figure 2) to evaluate the influence of the KG embedding on the semantic similarity:
• KG Embedding Retrieval Module 1: it takes into consideration the entity embedding index
and follows the same principle utilized in the Wikipedia Knowledge Retrieval Module. For a
given sentence, the top-k semantically similar documents are retrieved over the sentence
embedding field.
• KG Embedding Retrieval Module 2: it retrieves the top-1 semantically similar document.
        </p>
        <p>Then, a second search over the entity dense vector index is performed to retrieve the
top-k similar documents based on the KG embeddings of the entities.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Named Entity Recognition and Classification</title>
      <p>
        In CLEF-HIPE-2022 [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ], the named entity recognition and classification (NERC) task consists
in the recognition and classification of entities, such as people and locations, within historical
7https://www.wikidata.org/
8https://graphvite.io/docs/latest/pretrained_model.html
multilingual newspapers and classical commentaries. According to the organizers [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ], it is
composed of two sub-tasks with diferent levels of dificulty:
• Subtask 1.1 - NERC-Coarse: the identification and categorization of entity mentions
according to high-level entity types (e.g., Person, Location).
• Subtask 1.2 - NERC-Fine: the recognition and classification of entity mentions at diferent
levels, finer-grained entity types and nested entities, up to one level of depth (nested
entities).
      </p>
      <sec id="sec-4-1">
        <title>4.1. NERC Architecture</title>
        <p>
          Our proposed architecture is presented in Figure 3. In the right, we detail our model that
consists in a Base Model with new adapter layers, and the encoding of the additional contexts
(context jokers). As an overview, after the contexts are generated for an initial sentence, we
encode the tokens of the sentence with the Base Model, while the additional contexts are only
passed through the BERT pre-trained model encoder. These representations are afterward
concatenated, followed by the prediction CRF-based layers. In the left, we present two example
sentences from hipe-2020 and ajmc, for demonstrating the diferent levels of the entity types.
Base Model Our base model is based on the architecture proposed for CLEF-HIPE-2020
[
          <xref ref-type="bibr" rid="ref13 ref25">13, 25</xref>
          ] that consists in a hierarchical, multitask learning approach, with a fine-tuned encoder
based on BERT. The previous model included an encoder with two Transformer [
          <xref ref-type="bibr" rid="ref21">21</xref>
          ] layers
on top of the BERT pre-trained model encoder. This year, we add adapter modules to these
layers [
          <xref ref-type="bibr" rid="ref34">34</xref>
          ]. The adapters are added to each Transformer layer after the projection following
multi-headed attention. The adapter consists of a bottleneck which contains few parameters
relative to the attention and feed-forward layers in the original model. This acts as a task
adapter [
          <xref ref-type="bibr" rid="ref24">24</xref>
          ] for fine-grained NER. The attention modules in the Transformer layers adapt not
only to the task, but also to the noisy input which proved to increase performance of NER in
such special conditions [
          <xref ref-type="bibr" rid="ref25">25</xref>
          ]. Finally, the multitask prediction layer consists of separate CRF
layers9.
        </p>
        <p>Context Jokers In order to include the additional contexts generated as explained in
Section 3, we introduce the context jokers. Each additional context is passed through the BERT
pre-trained model encoder10 which is afterward mean-pooled along the sequence axis11. We
call this representation the context joker. The context jokers are afterward concatenated with
the sequential representation of the initial tokens of the sentence, as seen in Figure 3 and they
are discarded at the moment of prediction. We call them jokers because we see them as wild
cards unobtrusively inserted in the representation of the current sentence for improving the
recognition of the fine-grained entities. However, we also consider that these jokers can afect
the results in a way not immediately apparent and could also be harmful to the performance of
a NERC system.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Experiments and Internal Results</title>
        <p>
          CLEF-HIPE-2022 consists in assessing both tasks, NERC and EL in terms of precision (P),
recall (R), and F-measure (F1) at macro and micro levels [
          <xref ref-type="bibr" rid="ref10 ref35">35, 10</xref>
          ]12. Two evaluation scenarios
are considered: strict (exact boundary matching) and fuzzy boundary matching. For our
internal NERC results, we report only the strict matching (NERC-Coarse and NERC-Fine).
Our experimental setup consists in a baseline model and three settings with diferent levels of
knowledge-based contexts:
• no-context: Base Model with bert-base-multilingual-cased13;
• v0-language-specific: context jokers are generated with Wikipedia Knowledge
Retrieval Module;
• v1-en-wk5m: context jokers are generated with KG Embedding Retrieval Module 1;
• v2-en-wk5m: contexts jokers are generated with KG Embedding Retrieval Module 2.
French Our preliminary results for French, hipe-2020 and ajmc datasets, are shown in
Table 2. They reveal that generating contexts with KG Embedding Retrieval Module 1 &amp; 2 brings
considerable improvements for HIPE even if our Base Model provides the higher precision
for NERC-Coarse and the Wikipedia Knowledge Retrieval Module the higher recall for both
granularities. Adding any type of context to ajmc seems to slightly afect the precision while
9There is a CRF layer for each level of the entity types (NE-COARSE-LIT, NE-COARSE-METO, NE-FINE-LIT,
NE-FINE-METO, NE-FINE-COMP, NE-NESTED), thus six layers. If a dataset does not have fine-grained entities (e.g.,
English in hipe-2020, we maintain the same numbers of layers, and the model will learn to predict no entity.
        </p>
        <p>
          10We do not utilize in this case the additional Transformer layers with adapters, since these were specifically
proposed for noisy text and they do not bring any increase in performance as observed by Boroş et al. [
          <xref ref-type="bibr" rid="ref25">25</xref>
          ].
        </p>
        <p>11The maximum length of each context corresponds to the one handled by the language model. Thus, for example,
for a BERT-base model, the maximum is 512.</p>
        <p>12We utilized the HIPE-scorer https://github.com/hipe-eval/HIPE-scorer.</p>
        <p>13https://huggingface.co/bert-base-multilingual-cased
the contexts produced by the KG Embedding Retrieval Module 2 has a positive impact for the
NERC-Coarse recall.</p>
        <p>German As for German, our preliminary results presented in Table 3 show the larger
improvements when applying contexts for both ajmc and hipe-2020, specially with KG Embedding
Retrieval Module 1 &amp; 2. We assume that this is due the considerably smaller training dataset
than for the other languages.</p>
        <p>English Our preliminary results for English, shown in Table 4, indicate that generating
contexts with KG Embedding Retrieval Module 1 &amp; 2 brings considerable improvements on ajmc
for both granularities. Adding contexts to hipe-2020 has a double efect. They negatively
impact precision while improving recall. This is due to the lack of English training documents
and the fact that the contexts were generated using the French and German hipe-2020 training
datasets14.</p>
      </sec>
      <sec id="sec-4-3">
        <title>4.3. CLEF-HIPE-2022 Results</title>
        <p>The oficial CLEF-HIPE-2022 competition was restricted to two submissions. We, thus, selected
our baseline (no-context) and our best context generator models (v2-en-wk5m). In order to
improve the performance of our models, we stacked, for each language, a language-specific
language model. For English, we add bert-base-cased15, while for French and German,
we add the open-source French and German Europeana BERT models pretrained on the open
source Europeana digitized newspapers provided by The European Library and published by
the MDZ Digital Library team (dbmdz)16.</p>
        <p>French, German, English Our oficial results for French, German and English are shown in
Tables 5, 6, and 7 respectively. Adding contexts with the KG Embedding Retrieval Module 2 reveals
14These training sets were combined and used for training the model. Since the English hipe-2020 has only
NERC-Coarse entities, we discarded the NERC-Fine and the nested entities from the the French and German
hipe-2020, before training.</p>
        <p>15We utilized the English BERT model https://huggingface.co/bert-base-cased.</p>
        <p>16We utilized the bert-base-french-europeana-cased and bert-base-german-europeana-cased
from https://huggingface.co/dbmdz/.
a general improvement for all languages for ajmc. The additional contexts for hipe-2020
behave diferently. For French, our baseline model performed better for coarse granularity with
exact boundary matching. For German, contexts improved performance for coarse granularity
while slightly negatively afecting fine granularity. Finally, for English, the KG Embedding
Retrieval Module 2 boosted the performance for the coarse-grained entities.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Entity Linking</title>
      <p>In CLEF-HIPE-2022, the EL task consists in the disambiguation of named entities using two
settings:
• EL-only: The ground-truth regarding the entity mentions is provided, hence the entity
disambiguation runs uses the gold entity mentions of NERC and the only variable is the
EL system;
• End-to-end EL: No prior knowledge of the named entities is given, therefore EL has to
be performed over the named entities predicted with the NERC models (no-context
and v2-en-wk5m).</p>
      <p>
        Our EL model is based on the same neural approach that we proposed for CLEF-HIPE-2020 [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ].
It is combined with a filtering process to analyze the historical mentions and to disambiguate
them using the Wikidata KB [36]. Combining information from Wikipedia, Wikidata, and
DBpedia allows a thorough analysis of the characteristics of the entities and, as in
CLEF-HIPE2020, it helped our method to correctly disambiguate mentions in historical documents. Table 8
presents our EL scores for CLEF-HIPE-2022 in terms of P, R, and F1 for the hipe-2020 dataset.
It can be observed that adding contexts to German and English has a negative impact on the
recall which is consistent with our NERC results (cf. Table 6 and Table 7). Results also show
that applying additional contexts to French does not increase performances. The extended
results and ranking of CLEF-HIPE-2022 are available at the oficial website of the evaluation
campaign17.
      </p>
    </sec>
    <sec id="sec-6">
      <title>6. Conclusions</title>
      <p>For the participation of our team (L3i) in CLEF-HIPE-2022, we proposed two neural-based
methods for the tasks of NERC and EL. We conclude, for NERC, that our joker-based approach
generally performed well, due to the additional KG-based contexts and model improvements in
regards to the treatment of such contexts. For EL, the model we proposed for CLEF-HIPE-2020
confirmed its good performance, with and without context. Finally, we consider that external
knowledge has brought clear improvements to both our approaches and future work on this
subject could furthermore prove the utility and importance of high-quality symbolic knowledge.</p>
    </sec>
    <sec id="sec-7">
      <title>Acknowledgments</title>
      <p>This work has been supported by the ANNA (2019-1R40226) and TERMITRAD
(2020-20198510010) projects funded by the Nouvelle-Aquitaine Region, France. We would like to also
thank Nicolas Sidère and Jean-Loup Guillaume for the insightful discussions.
[36] E. Linhares Pontes, L. A. Cabrera-Diego, J. G. Moreno, E. Boros, A. Hamdi, A. Doucet,
N. Sidere, M. Coustaty, Melhissa: a multilingual entity linking architecture for historical
press articles, International Journal on Digital Libraries 23 (2022) 133–160.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>F.</given-names>
            <surname>Boschetti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Cimino</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Dell'Orletta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Lebani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Passaro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Picchi</surname>
          </string-name>
          , G. Venturi,
          <string-name>
            <given-names>S.</given-names>
            <surname>Montemagni</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Lenci</surname>
          </string-name>
          ,
          <article-title>Computational analysis of historical documents: An application to italian war bulletins in world war i and ii</article-title>
          , in: Workshop on Language resources
          <article-title>and technologies for processing and linking historical documents and archives</article-title>
          (LRT4HDA
          <year>2014</year>
          ), ELRA,
          <year>2014</year>
          , pp.
          <fpage>70</fpage>
          -
          <lpage>75</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>M.</given-names>
            <surname>Rovera</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Nanni</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. P.</given-names>
            <surname>Ponzetto</surname>
          </string-name>
          ,
          <article-title>Providing advanced access to historical war memoirs through the identification of events, participants and roles (</article-title>
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>A.</given-names>
            <surname>Cybulska</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Vossen</surname>
          </string-name>
          ,
          <article-title>Historical event extraction from text</article-title>
          ,
          <source>in: Proceedings of the 5th ACL-HLT Workshop on Language Technology for Cultural Heritage</source>
          ,
          <source>Social Sciences, and Humanities</source>
          ,
          <year>2011</year>
          , pp.
          <fpage>39</fpage>
          -
          <lpage>43</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>E.</given-names>
            <surname>Boschee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Natarajan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Weischedel</surname>
          </string-name>
          ,
          <article-title>Automatic extraction of events from open source text for predictive forecasting</article-title>
          , in: Handbook of Computational Approaches to Counterterrorism, Springer,
          <year>2013</year>
          , pp.
          <fpage>51</fpage>
          -
          <lpage>67</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>S.</given-names>
            <surname>Oberbichler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Boroş</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Doucet</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Marjanen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Pfanzelter</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Rautiainen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Toivonen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Tolonen</surname>
          </string-name>
          ,
          <article-title>Integrated interdisciplinary workflows for research on historical newspapers: Perspectives from humanities scholars, computer scientists, and librarians</article-title>
          ,
          <source>Journal of the Association for Information Science and Technology</source>
          <volume>73</volume>
          (
          <year>2022</year>
          )
          <fpage>225</fpage>
          -
          <lpage>239</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>S.</given-names>
            <surname>Hechl</surname>
          </string-name>
          , P.-C. Langlais,
          <string-name>
            <given-names>J.</given-names>
            <surname>Marjanen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Oberbichler</surname>
          </string-name>
          , E. Pfanzelter,
          <article-title>Digital interfaces of historical newspapers: opportunities, restrictions and recommendations</article-title>
          ,
          <source>Journal of Data Mining &amp; Digital Humanities</source>
          (
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>S.</given-names>
            <surname>Oberbichler</surname>
          </string-name>
          , E. Pfanzelter,
          <article-title>Topic-specific corpus building: A step towards a representative newspaper corpus on the topic of return migration using text mining methods</article-title>
          ,
          <source>Journal of Digital History</source>
          <volume>1</volume>
          (
          <year>2021</year>
          )
          <fpage>74</fpage>
          -
          <lpage>98</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>M.</given-names>
            <surname>Ehrmann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Hamdi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E. Linhares</given-names>
            <surname>Pontes</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Romanello</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Douvet</surname>
          </string-name>
          ,
          <article-title>A Survey of Named Entity Recognition and Classification in Historical Documents, ACM Computing Surveys (2022 (to appear))</article-title>
          . URL: https://arxiv.org/abs/2109.11406.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>E.</given-names>
            <surname>Linhares Pontes</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. A.</given-names>
            <surname>Cabrera-Diego</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. G.</given-names>
            <surname>Moreno</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Boros</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Hamdi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Sidère</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Coustaty</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Doucet</surname>
          </string-name>
          ,
          <article-title>Entity linking for historical documents: challenges and solutions</article-title>
          ,
          <source>in: International Conference on Asian Digital Libraries</source>
          , Springer,
          <year>2020</year>
          , pp.
          <fpage>215</fpage>
          -
          <lpage>231</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>M.</given-names>
            <surname>Ehrmann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Romanello</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Bircher</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Clematide</surname>
          </string-name>
          ,
          <article-title>Introducing the CLEF 2020 HIPE shared task: Named entity recognition and linking on historical newspapers</article-title>
          , in: J. M. Jose, E. Yilmaz,
          <string-name>
            <given-names>J.</given-names>
            <surname>Magalhães</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Castells</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Ferro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. J.</given-names>
            <surname>Silva</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Martins</surname>
          </string-name>
          (Eds.),
          <source>Advances in information retrieval</source>
          , Springer International Publishing, Cham,
          <year>2020</year>
          , pp.
          <fpage>524</fpage>
          -
          <lpage>532</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>M.</given-names>
            <surname>Ehrmann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Romanello</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Flückiger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Clematide</surname>
          </string-name>
          , Overview of clef hipe 2020:
          <article-title>Named entity recognition and linking on historical newspapers</article-title>
          ,
          <source>in: International Conference of the Cross-Language Evaluation Forum for European Languages</source>
          , Springer,
          <year>2020</year>
          , pp.
          <fpage>288</fpage>
          -
          <lpage>310</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>M.</given-names>
            <surname>Ehrmann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Romanello</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Flückiger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Clematide</surname>
          </string-name>
          ,
          <article-title>Extended overview of clef hipe 2020: named entity processing on historical newspapers</article-title>
          ,
          <source>in: CEUR Workshop Proceedings</source>
          , 2696,
          <string-name>
            <surname>CEUR-WS</surname>
          </string-name>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>E.</given-names>
            <surname>Boros</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E. L.</given-names>
            <surname>Pontes</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. A.</given-names>
            <surname>Cabrera-Diego</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Hamdi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. G.</given-names>
            <surname>Moreno</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Sidère</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Doucet</surname>
          </string-name>
          ,
          <article-title>Robust named entity recognition and linking on historical multilingual documents</article-title>
          , in: Conference and
          <article-title>Labs of the Evaluation Forum (CLEF</article-title>
          <year>2020</year>
          ), volume
          <volume>2696</volume>
          ,
          <string-name>
            <surname>CEUR-WS Working</surname>
            <given-names>Notes</given-names>
          </string-name>
          ,
          <year>2020</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>17</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>K.</given-names>
            <surname>Labusch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Neudecker</surname>
          </string-name>
          ,
          <article-title>Named entity disambiguation and linking historic newspaper ocr with bert</article-title>
          .,
          <source>in: CLEF (Working Notes)</source>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>S.</given-names>
            <surname>Schweter</surname>
          </string-name>
          , L. März,
          <article-title>Triple e-efective ensembling of embeddings and language models for ner of historical german</article-title>
          .,
          <source>in: CLEF (Working Notes)</source>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>V.</given-names>
            <surname>Provatorova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Vakulenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Kanoulas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Dercksen</surname>
          </string-name>
          ,
          <string-name>
            <surname>J. M. van Hulst</surname>
          </string-name>
          ,
          <article-title>Named entity recognition and linking on historical newspapers: Uva. ilps &amp; rel at clef hipe 2020 (</article-title>
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>T.</given-names>
            <surname>Kristanti</surname>
          </string-name>
          , L. Romary,
          <article-title>Delft and entity-fishing: Tools for clef hipe 2020 shared task</article-title>
          ,
          <source>in: CLEF 2020-Conference and Labs of the Evaluation Forum</source>
          , volume
          <volume>2696</volume>
          ,
          <string-name>
            <surname>CEUR</surname>
          </string-name>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <surname>C. B. El Vaigh</surname>
            ,
            <given-names>G. Le</given-names>
          </string-name>
          <string-name>
            <surname>Noé-Bienvenu</surname>
            , G. Gravier,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Sébillot</surname>
          </string-name>
          ,
          <article-title>Irisa system for entity detection and linking at clef hipe 2020</article-title>
          , in: CEUR Workshop Proceedings,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>P. J. O.</given-names>
            <surname>Suárez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Dupont</surname>
          </string-name>
          , G. Lejeune, T. Tian, Sinner@
          <fpage>clef</fpage>
          -hipe2020:
          <article-title>Sinful adaptation of sota models for named entity recognition in french and german</article-title>
          ,
          <source>in: CLEF (Working Notes)</source>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>M.</given-names>
            <surname>Ehrmann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Romanello</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Doucet</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Clematide</surname>
          </string-name>
          ,
          <article-title>Introducing the hipe 2022 shared task: Named entity recognition and linking in multilingual historical documents</article-title>
          ,
          <source>in: European Conference on Information Retrieval</source>
          , Springer,
          <year>2022</year>
          , pp.
          <fpage>347</fpage>
          -
          <lpage>354</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>A.</given-names>
            <surname>Vaswani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Shazeer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Parmar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Uszkoreit</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Jones</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. N.</given-names>
            <surname>Gomez</surname>
          </string-name>
          , Ł. Kaiser,
          <string-name>
            <surname>I. Polosukhin</surname>
          </string-name>
          ,
          <article-title>Attention is all you need</article-title>
          ,
          <source>Advances in neural information processing systems</source>
          <volume>30</volume>
          (
          <year>2017</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>N.</given-names>
            <surname>Kolitsas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.-E.</given-names>
            <surname>Ganea</surname>
          </string-name>
          , T. Hofmann,
          <article-title>End-to-end neural entity linking</article-title>
          ,
          <source>in: Proceedings of the 22nd Conference on Computational Natural Language Learning</source>
          ,
          <year>2018</year>
          , pp.
          <fpage>519</fpage>
          -
          <lpage>529</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>S.-A.</given-names>
            <surname>Rebufi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Bilen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Vedaldi</surname>
          </string-name>
          ,
          <article-title>Learning multiple visual domains with residual adapters</article-title>
          ,
          <source>Advances in neural information processing systems</source>
          <volume>30</volume>
          (
          <year>2017</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>J.</given-names>
            <surname>Pfeifer</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Vulić</surname>
          </string-name>
          , I. Gurevych, S. Ruder,
          <article-title>MAD-X: An Adapter-Based Framework for MultiTask Cross-Lingual Transfer</article-title>
          ,
          <source>in: Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP)</source>
          ,
          <article-title>Association for Computational Linguistics</article-title>
          , Online,
          <year>2020</year>
          , pp.
          <fpage>7654</fpage>
          -
          <lpage>7673</lpage>
          . URL: https://aclanthology.org/
          <year>2020</year>
          .emnlp-main.
          <volume>617</volume>
          . doi:
          <volume>10</volume>
          . 18653/v1/
          <year>2020</year>
          .emnlp-main.
          <volume>617</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>E.</given-names>
            <surname>Boroş</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Hamdi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E. L.</given-names>
            <surname>Pontes</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.-A.</given-names>
            <surname>Cabrera-Diego</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. G.</given-names>
            <surname>Moreno</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Sidere</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Doucet</surname>
          </string-name>
          ,
          <article-title>Alleviating digitization errors in named entity recognition for historical documents</article-title>
          ,
          <source>in: Proceedings of the 24th conference on computational natural language learning</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>431</fpage>
          -
          <lpage>441</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [26]
          <string-name>
            <given-names>M.</given-names>
            <surname>Romanello</surname>
          </string-name>
          , S. Najem-Meyer,
          <string-name>
            <given-names>B.</given-names>
            <surname>Robertson</surname>
          </string-name>
          ,
          <article-title>Optical character recognition of 19th century classical commentaries: the current state of afairs</article-title>
          ,
          <source>in: The 6th International Workshop on Historical Document Imaging and Processing</source>
          ,
          <year>2021</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>6</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [27]
          <string-name>
            <given-names>S.</given-names>
            <surname>Mayhew</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Tsygankova</surname>
          </string-name>
          ,
          <string-name>
            <surname>D.</surname>
          </string-name>
          <article-title>Roth, ner and pos when nothing is capitalized</article-title>
          ,
          <source>in: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP)</source>
          ,
          <article-title>Association for Computational Linguistics</article-title>
          , Hong Kong, China,
          <year>2019</year>
          , pp.
          <fpage>6256</fpage>
          -
          <lpage>6261</lpage>
          . URL: https://aclanthology.org/D19-1650. doi:
          <volume>10</volume>
          .18653/v1/
          <fpage>D19</fpage>
          -1650.
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          [28]
          <string-name>
            <given-names>X.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Shen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Cai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Xie</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Lu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhuang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Tu</surname>
          </string-name>
          , et al.,
          <article-title>Damo-nlp at semeval-2022 task 11: A knowledge-based system for multilingual named entity recognition</article-title>
          ,
          <source>arXiv preprint arXiv:2203.00545</source>
          (
          <year>2022</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          [29]
          <string-name>
            <given-names>X.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Gao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Tang</surname>
          </string-name>
          ,
          <article-title>Kepler: A unified model for knowledge embedding and pre-trained language representation</article-title>
          , arXiv preprint arXiv:
          <year>1911</year>
          .
          <volume>06136</volume>
          (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          [30]
          <string-name>
            <given-names>I.</given-names>
            <surname>Yamada</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Asai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Sakuma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Shindo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Takeda</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Takefuji</surname>
          </string-name>
          ,
          <string-name>
            <surname>Y. Matsumoto,</surname>
          </string-name>
          <article-title>Wikipedia2Vec: An eficient toolkit for learning and visualizing the embeddings of words and entities from Wikipedia</article-title>
          ,
          <source>in: Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, Association for Computational Linguistics</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>23</fpage>
          -
          <lpage>30</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          [31]
          <string-name>
            <given-names>N.</given-names>
            <surname>Reimers</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Gurevych</surname>
          </string-name>
          , Sentence-BERT:
          <article-title>Sentence embeddings using Siamese BERTnetworks</article-title>
          ,
          <source>in: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP)</source>
          ,
          <article-title>Association for Computational Linguistics</article-title>
          , Hong Kong, China,
          <year>2019</year>
          , pp.
          <fpage>3982</fpage>
          -
          <lpage>3992</lpage>
          . URL: https://aclanthology.org/D19-1410. doi:
          <volume>10</volume>
          .18653/v1/
          <fpage>D19</fpage>
          -1410.
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          [32]
          <string-name>
            <given-names>N.</given-names>
            <surname>Reimers</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Gurevych</surname>
          </string-name>
          ,
          <article-title>Making monolingual sentence embeddings multilingual using knowledge distillation</article-title>
          ,
          <source>in: Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP)</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>4512</fpage>
          -
          <lpage>4525</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref33">
        <mixed-citation>
          [33]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.-H.</given-names>
            <surname>Deng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.-Y.</given-names>
            <surname>Nie</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Tang</surname>
          </string-name>
          ,
          <article-title>Rotate: Knowledge graph embedding by relational rotation in complex space</article-title>
          , arXiv preprint arXiv:
          <year>1902</year>
          .
          <volume>10197</volume>
          (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref34">
        <mixed-citation>
          [34]
          <string-name>
            <given-names>N.</given-names>
            <surname>Houlsby</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Giurgiu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Jastrzebski</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Morrone</surname>
          </string-name>
          ,
          <string-name>
            <surname>Q. De Laroussilhe</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Gesmundo</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Attariyan</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Gelly</surname>
          </string-name>
          ,
          <article-title>Parameter-eficient transfer learning for nlp</article-title>
          ,
          <source>in: International Conference on Machine Learning, PMLR</source>
          ,
          <year>2019</year>
          , pp.
          <fpage>2790</fpage>
          -
          <lpage>2799</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref35">
        <mixed-citation>
          [35]
          <string-name>
            <given-names>J.</given-names>
            <surname>Makhoul</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Kubala</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Schwartz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Weischedel</surname>
          </string-name>
          , et al.,
          <article-title>Performance measures for information extraction</article-title>
          ,
          <source>in: Proceedings of DARPA broadcast news workshop</source>
          , Herndon,
          <string-name>
            <surname>VA</surname>
          </string-name>
          ,
          <year>1999</year>
          , pp.
          <fpage>249</fpage>
          -
          <lpage>252</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>