<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>Train</journal-title>
      </journal-title-group>
      <issn pub-type="ppub">1613-0073</issn>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>with Triple-aware Pre-Training</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Xi Yan</string-name>
          <email>xi.yan@uni-hamburg.de</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Cedric Möller</string-name>
          <email>cedric.moeller@uni-hamburg.de</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ricardo Usbeck</string-name>
          <email>ricardo.usbeck@leuphana.de</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Leuphana Universität Lüneburg</institution>
          ,
          <addr-line>Universitätsallee 1, 21335 Lüneburg</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Universität Hamburg</institution>
          ,
          <addr-line>Edmund-Siemers-Allee 1, 20146, Hamburg</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
      </contrib-group>
      <volume>5</volume>
      <issue>784</issue>
      <abstract>
        <p>The large-scale analysis of scientific and technical documents is crucial for extracting structured knowledge from unstructured text. A key challenge in this process is linking biomedical entities, as these entities are sparsely distributed and often underrepresented in the training data of large language models (LLM). At the same time, those LLMs are not aware of high level semantic connection between diferent biomedical entities, which are useful in identifying similar concepts in diferent textual contexts. To cope with aforementioned problems, some recent works focused on injecting knowledge graph information into LLMs. However, former methods either ignore the relational knowledge of the entities or lead to catastrophic forgetting. Therefore, we propose a novel framework to pre-train the powerful generative LLM by a corpus synthesized from a KG. In the evaluations we are unable to confirm the benefit of including synonym, description or relational information. This work-in-progress highlights key challenges and invites further discussion on leveraging semantic information for LLm performance and on scientific document processing.</p>
      </abstract>
      <kwd-group>
        <kwd>Entity Linking</kwd>
        <kwd>Scientific data</kwd>
        <kwd>Deep learning</kwd>
        <kwd>Semantic information</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Biomedical entity linking (EL) is a critical process in biomedical text mining that seeks to identify and
associate relevant biological and medical entities mentioned in unstructured text with their
corresponding identifiers in knowledge bases. EL systems have also been combined to promote the knowledge
acquisition task[
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. Accurate recognition and linking of these entities are pivotal in promoting
biomedical research, drug discovery, and personalized medicine [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. Although substantial progress has been
made in recent years, there is an ongoing need for refining methods and techniques employed for entity
linking in the biomedical domain.
      </p>
      <p>
        In this report, we present a novel approach that integrates linearized (in which a graph is traversed
and encoded when producing the linearized representationHoyle et al. [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ].) triples into the biomedical
entity linking process while reevaluating the inclusion of synonym information. Our proposed method
linearizes triples and considers them during the pre-training step. In past studies, synonym information,
which involves using alternative names or terminologies for the same biomedical entity, has been
proven to enhance entity linking when used during pre-training [
        <xref ref-type="bibr" rid="ref4 ref5">4, 5</xref>
        ]. Our study aims to build upon
this existing knowledge by integrating both strategies and assessing their impact on performance.
      </p>
      <p>
        Despite the reported benefits of synonym information in prior studies, our analysis of this approach,
combined with the introduction of linearized triples [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], yielded diferent results. We find that
incorporating linearized triples only lead to minimal improvements in our entity linking model’s performance.
Moreover, we are unable to confirm the purported advantages of including synonym information in
our experiments, which stands in contrast to the findings of previous literature.
      </p>
      <p>We highlight the limitations of our study and suggest possible avenues for future research to further
advance biomedical entity linking techniques by building on our work with linearized triples and</p>
      <p>https://www.hcds.uni-hamburg.de/hcds/head-hcds/xi-yan.html (X. Yan);
https://www.hcds.uni-hamburg.de/hcds/head-hcds/cedric-moeller.html (C. Möller);</p>
      <p>CEUR</p>
      <p>ceur-ws.org
reevaluating synonym information. The code is available at our GitHub repo 1.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related work</title>
      <p>
        Entity Linking has a long history of research. Recent methods can be categorized into two types. First,
discriminative methods that are based on the bi-encoder / cross-encoder pairing [
        <xref ref-type="bibr" rid="ref7 ref8 ref9">7, 8, 9</xref>
        ]. Both encoders
are commonly BERT-like models. The bi-encoder encodes the description of each entity and matches it
to the text by using an approximate nearest neighbor search. This is important as the next step, the
cross-encoding, is expensive. Here, those neighbors are reranked by applying a cross-encoder to the
concatenation of both, the input text and the entity description. The highest-ranked entity is then the
ifnal linked one. In the biomedical domain, the works by [ 10], [11], [12] and [13] fall into this category.
      </p>
      <p>
        Another type of entity linker is based on generative models [
        <xref ref-type="bibr" rid="ref5">14, 15, 5</xref>
        ]. Here, instead of using some
external description of an entity, the whole model memorizes the KG during training. The linked entity
is then directly generated by the model. Such methods skip the problem of mining negatives which are
crucial for a good performance of bi-encoder-based methods. BioLinkerAI [16] and Gallego et al [17]
use the entity definitions and therasaurs (i.e., UMLS) to enhance the performance of LLM. Only the
work by Yuan et al. [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] is based on such methods in the biomedical domain. As generative models lack
the ability to incorporate external information, they alleviate this problem by introducing a pre-training
stage where syntactical information from a knowledge graph is learned. This is especially important in
the biomedical domain as entities often own a large variety of synonyms. We build upon their work by
extending the pre-training regime to the inclusion of triple information.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Method</title>
      <sec id="sec-3-1">
        <title>3.1. Task definition</title>
        <p>
          Given are a text  , a set of marked mentions   in the text and a KG  = (ℰ , ℛ, ) . The KG consists
of a set of entities ℰ, a set of relations ℛ and a set of edges composed of head entity, relation and tail
entity  ⊆ (ℰ × ℛ × ℰ ) . The task is to identify the subset of entities   ⊆ ℰ which the mentions   are
referring to.
3.2. Model
In the vein of the work by [14], we model the problem as a sequence-to-sequence generation task.
The input to the generative model is text and the output are the generated entity identifiers in the
corresponding KGs. Similar to other works [
          <xref ref-type="bibr" rid="ref4">14, 15, 4</xref>
          ], we consider the definition of the concepts in the
corresponding KGs as the unique textual representation of each concept. The definition and synonyms
are short and unique, and will not introduce the problem of ambiguation of entities.
        </p>
      </sec>
      <sec id="sec-3-2">
        <title>3.3. Pre-training</title>
        <p>
          We linearize the information from synonym and triples in the pre-training stage. An overview of
the pre-training and an example is give in Figure 1 . They are linearized into a synthesized corpora
before feeding into the BART. We have tested 2 diferent settings for converting the triples, namely
line-by-line and all-in-one. We add triple pre-training step on which add the triples information to
the LLM, on top of synonym, which is used by [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ].
        </p>
        <p>
          In terms of the synonym information, we follow the setting by [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ]. We first extract the description
of the entity and convert it to a text of the following form:
[BOS][ST]   [ET] is defined as   [EOS]
(1)
This lets the model learn the connection between the diferent synonyms of the same entity.
        </p>
        <p>Based on that, we introduce an additional pre-training step to incorporate more semantic information
by utilising triple information from the underlying knowledge graph. A triple is of the form &lt; ,  ,  ′ &gt;
which describes that a relationship  holds between entity  and  ′. The input is here the same as for
the synonym information. The output is of the form:</p>
        <p>is here the label of relation  . We denote this line-by-line. Furthermore, we experimented with an
all-in-one pre-training approach of the form:
[BOS]</p>
        <p>′ [EOS]
[BOS]  
 
 1   1
 1 …  

 

  [EOS]
Here,   stands for the synonym  and   for the description of entity  . This would be the input to the
encoder of the generative model.</p>
        <p>As an output the model has to generate:
(2)
(3)
(4)
(5)</p>
      </sec>
      <sec id="sec-3-3">
        <title>3.4. Fine-tuning</title>
        <p>
          template as follows:
During fine-tuning, the model is trained for the actual entity linking task. The input to the generative
model is the unlabelled biomedical text. To generate the linked entities, each mention is included in a
[BOS]   is   [EOS]
The model then generates the entity identifier after the token ”
is”. Similar to the work by Yuan et al.
[
          <xref ref-type="bibr" rid="ref4">4</xref>
          ], we choose the synonym which is syntactically close to the corresponding mention in the text as
the target entity identifier during fine-tuning.
        </p>
        <p>The generated entity identifier is mapped back to the concrete entity in the final step via a lookup</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Evaluation</title>
      <sec id="sec-4-1">
        <title>4.1. Pre-training Strategy</title>
        <p>We use a synthesized corpus composed of triples, synonyms and descriptions from UMLS. More
specifically, we decide to use a subset of UMLS, st21pv [ 18]. It is a well-connected KG with information
about concept definitions and synonyms. Specifically, 160K out of 2.37M concepts have definitions,
1.11M concepts have several synonyms and 68K concepts are connected to on average 8 triples as a
subject in a single hop. During the pre-training step, we construct samples by iterating through each
concept’s synonyms and triples. Each concept is densely connected and the distribution of the number
of triples a concept is connected to is skewed. For instance, some ”popular” concepts are connected to
over 1000 triples, while some are connected to only 1 triple. To avoid the class imbalance, we sample
the included triples based on the relation frequencies.</p>
        <p>To train the model with KG information, we linearize triples. Linearization refers to a special type of
technique on converting graph to text, i.e., converting triples to one/more sentences which serve as
input of the LLM.</p>
        <p>We sample the included triples based on the relation frequencies. First, we gather the occurrence
frequency of all relations in the KB by counting the number of triples this relation is connected to.</p>
        <p>
          Both settings are trained under the same experiment setting with a batch size of 128. We save the best
model within 12 training epochs. We experiment with BART-base, bioBART-Large, and bioBART-Base.
We choose BART to align to the work of [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ] so that we can make comparison about whether relational
information is beneficial to the model. Note that we define the probability (   ) of a relation  to be
negatively related to the frequency. Then, for each concept in the KG, we collect its connected triples
and segment the triples into diferent groups based on their relation  .
4.1.1. Fine-tuning
The model is fine-tuned on two established datasets, namely BC5CDR [ 19] and NCBI [20]. Those entity
linking datasets are constructed on subsets of UMLS, making them perfect choices to test our model’s
performance on. Among the datasets, NCBI and BC5CDR are generated by annotating PubMed papers.
On the other hand, NCBI and BC5CDR are annotated against Medical Subject Headings (MeSH) - a
terminology knowledge graph for indexing and cataloging of biomedical information.
        </p>
        <p>The statistics of the four datasets are exhibited in Table 1 below. As we can see, NCBI and BC5CDR
(annotated on academic text) are smaller in size. Also NCBI and BC5CDR are dense in terms of the
target entities they contain (14,967 and 268,162).</p>
        <p>BART-large [21] is chosen as the generative model as it has been an established benchmark model
for such tasks.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Results</title>
        <p>
          We assess the performance of four distinct models in the entity linking task, including two of our
own models, each pre-trained via either a line-by-line or all-in-one strategy, a synonym pre-trained
model from [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ] (denoted Syn-Only), and a basic BART model. We also include the recent papers which
pretrains BART on biomedical domain [22] before finetuned on biomedical entity linking datasets and
ResCNN [23] which achieves state-of-the-art results on various biomedical EL datasets. Each model
undergoes fine-tuning specific to the entity linking task. Recall@1 for each model are presented in the
Table 4.2. We limit ourselves to Recall@1 to follow the common practice when measuring entity linking
performance without named entity recognition. The best-performing metrics are emphasized in bold.
        </p>
      </sec>
      <sec id="sec-4-3">
        <title>4.3. Analysis</title>
        <p>Based on the table 4.2, our triple injection framework exceeds the BART baseline on the 2 benchmarks
datasets. On BC5CDR and NCBI, the gain compared to BART is around 0.2% and 0.5%.</p>
        <p>Does triple injection enhance model’s capacity to link to the correct entity? The answer is
yes, since over 2 datasets, the All-in-one or Line-by-line variants outperform the variant that was not
trained on the linearized corpora for around 1% (Recall@1).</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusion</title>
      <p>Our study seek to improve biomedical entity linking through the integration of linearized triples and
synonym information. However, contrary to expectation, the incorporation of these elements leads to
only minimal improvements in our EL model performance.</p>
      <p>In conclusion, our study underscores the complexities of biomedical EL and prompts the need for
more sophisticated approaches to improve its accuracy. A possible future extension of this work could
be to explore more sophisticated methods to instruct the LLMs to learn external knowledge, such that
the knowledge is injected in an eficient way which benefits the models in downstream tasks. For
instance, by incorporating the KG information not just in a linearized manner but by exploiting the
graph-structure with Graph Neural Networks [24], mutliple methods could be further developed.</p>
    </sec>
    <sec id="sec-6">
      <title>6. Acknowledgments</title>
      <p>This work has been partially supported by the Ministry of Research and Education within the project
‘RESCUE-MATE: Dynamische Lageerstellung und Unterstützung für Rettungskräfte in komplexen
Krisensituationen mittels Datenfusion und intelligenten Drohnenschwärmen’ (FKZ 13N16844), by the
Federal Ministry for Economic Afairs and Climate Action of Germany in the project CoyPu (project
number 01MK21007[G]). We utilized 2 x NVIDIA RTX A5000 24GB kindly provided by the NVIDIA
Academic Hardware Grant Program. The authors have no competing interests to declare that are
relevant to the content of this article.</p>
    </sec>
    <sec id="sec-7">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the author(s) used X-GPT-4 and Gramby in order to: Grammar
and spelling check. After using these tool(s)/service(s), the author(s) reviewed and edited the content as
needed and take(s) full responsibility for the publication’s content.
Computational Linguistics: Human Language Technologies: Industry Track, NAACL 2022, Hybrid:
Seattle, Washington, USA + Online, July 10-15, 2022, Association for Computational Linguistics,
2022, pp. 209–220. doi:10.18653/v1/2022.naacl- industry.24.
[10] R. Angell, N. Monath, S. Mohan, N. Yadav, A. McCallum, Clustering-based inference for
biomedical entity linking, in: K. Toutanova, A. Rumshisky, L. Zettlemoyer, D. Hakkani-Tür, I. Beltagy,
S. Bethard, R. Cotterell, T. Chakraborty, Y. Zhou (Eds.), Proceedings of the 2021 Conference of the
North American Chapter of the Association for Computational Linguistics: Human Language
Technologies, NAACL-HLT 2021, Online, June 6-11, 2021, Association for Computational Linguistics,
2021, pp. 2598–2608. doi:10.18653/v1/2021.naacl- main.205.
[11] M. Varma, L. J. Orr, S. Wu, M. Leszczynski, X. Ling, C. Ré, Cross-domain data integration for named
entity disambiguation in biomedical text, in: M. Moens, X. Huang, L. Specia, S. W. Yih (Eds.),
Findings of the Association for Computational Linguistics: EMNLP 2021, Virtual Event / Punta
Cana, Dominican Republic, 16-20 November, 2021, Association for Computational Linguistics,
2021, pp. 4566–4575. doi:10.18653/v1/2021.findings- emnlp.388.
[12] D. Agarwal, R. Angell, N. Monath, A. McCallum, Entity linking and discovery via
arborescencebased supervised clustering, CoRR abs/2109.01242 (2021). URL: https://arxiv.org/abs/2109.01242.
arXiv:2109.01242.
[13] R. Bhowmik, K. Stratos, G. de Melo, Fast and efective biomedical entity linking using a dual
encoder, in: E. Holderness, A. Jimeno-Yepes, A. Lavelli, A. Minard, J. Pustejovsky, F. Rinaldi (Eds.),
Proceedings of the 12th International Workshop on Health Text Mining and Information Analysis,
LOUHI@EACL, Online, April 19, 2021, Association for Computational Linguistics, 2021, pp. 28–37.</p>
      <p>URL: https://www.aclweb.org/anthology/2021.louhi-1.4/.
[14] N. D. Cao, G. Izacard, S. Riedel, F. Petroni, Autoregressive entity retrieval, in: 9th
International Conference on Learning Representations, ICLR 2021, Virtual Event, Austria, May 3-7, 2021,
OpenReview.net, 2021.
[15] N. D. Cao, L. Wu, K. Popat, M. Artetxe, N. Goyal, M. Plekhanov, L. Zettlemoyer, N. Cancedda,
S. Riedel, F. Petroni, Multilingual autoregressive entity linking, Trans. Assoc. Comput.
Linguistics 10 (2022) 274–290. URL: https://doi.org/10.1162/tacl_a_00460. doi:10.1162/tacl\_{a}{\_
{0}{0}{4}}{6}0.
[16] A. Sakor, K. Singh, M.-E. Vidal, Biolinkerai: Capturing knowledge using llms to enhance biomedical
entity linking, in: International Conference on Web Information Systems Engineering, Springer,
2024, pp. 262–272.
[17] F. Gallego, P. Ruas, F. M. Couto, F. J. Veredas, Enhancing cross-encoders using knowledge graph
hierarchy for medical entity linking in zero-and few-shot scenarios, Knowledge-Based Systems
314 (2025) 113211.
[18] S. Mohan, D. Li, Medmentions: A large biomedical corpus annotated with UMLS concepts, in: 1st
Conference on Automated Knowledge Base Construction, AKBC 2019, Amherst, MA, USA, May
20-22, 2019, 2019. URL: https://doi.org/10.24432/C5G59C. doi:10.24432/C5G59C.
[19] J. Li, Y. Sun, R. J. Johnson, D. Sciaky, C. Wei, R. Leaman, A. P. Davis, C. J. Mattingly, T. C.</p>
      <p>Wiegers, Z. Lu, Biocreative V CDR task corpus: a resource for chemical disease relation extraction,
Database J. Biol. Databases Curation 2016 (2016). URL: https://doi.org/10.1093/database/baw068.
doi:10.1093/database/baw068.
[20] R. I. Dogan, R. Leaman, Z. Lu, NCBI disease corpus: A resource for disease name recognition and
concept normalization, J. Biomed. Informatics 47 (2014) 1–10. URL: https://doi.org/10.1016/j.jbi.
2013.12.006. doi:10.1016/j.jbi.2013.12.006.
[21] P. Liu, W. Yuan, J. Fu, Z. Jiang, H. Hayashi, G. Neubig, Pre-train, prompt, and predict: A systematic
survey of prompting methods in natural language processing, ACM Computing Surveys 55 (2023)
1–35.
[22] H. Yuan, Z. Yuan, R. Gan, J. Zhang, Y. Xie, S. Yu, BioBART: Pretraining and evaluation of a
biomedical generative language model, in: Proceedings of the 21st Workshop on Biomedical
Language Processing, Association for Computational Linguistics, Dublin, Ireland, 2022, pp. 97–109.</p>
      <p>URL: https://aclanthology.org/2022.bionlp-1.9.
[23] T. Lai, H. Ji, C. Zhai, BERT might be overkill: A tiny but efective biomedical entity linker based
on residual convolutional neural networks, in: Findings of the Association for Computational
Linguistics: EMNLP 2021, Association for Computational Linguistics, Punta Cana, Dominican
Republic, 2021, pp. 1631–1639. URL: https://aclanthology.org/2021.findings-emnlp.140. doi:10.
18653/v1/2021.findings-emnlp.140.
[24] Z. Wu, S. Pan, F. Chen, G. Long, C. Zhang, S. Y. Philip, A comprehensive survey on graph neural
networks, IEEE transactions on neural networks and learning systems 32 (2020) 4–24.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>K.</given-names>
            <surname>Noullet</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Ourgani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Färber</surname>
          </string-name>
          ,
          <article-title>A full-fledged framework for combining entity linking systems and components</article-title>
          ,
          <source>in: Proceedings of the 12th Knowledge Capture Conference</source>
          <year>2023</year>
          , K-CAP '
          <fpage>23</fpage>
          ,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2023</year>
          , p.
          <fpage>148</fpage>
          -
          <lpage>156</lpage>
          . URL: https: //doi.org/10.1145/3587259.3627556. doi:
          <volume>10</volume>
          .1145/3587259.3627556.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>P.</given-names>
            <surname>Chandak</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Zitnik</surname>
          </string-name>
          ,
          <article-title>Building a knowledge graph to enable precision medicine</article-title>
          ,
          <source>Scientific Data</source>
          <volume>10</volume>
          (
          <year>2023</year>
          )
          <article-title>67</article-title>
          . doi:
          <volume>10</volume>
          .1038/s41597- 023- 01960- 3.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>A. M.</given-names>
            <surname>Hoyle</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Marasović</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N. A.</given-names>
            <surname>Smith</surname>
          </string-name>
          ,
          <article-title>Promoting graph awareness in linearized graph-totext generation</article-title>
          , in: C.
          <string-name>
            <surname>Zong</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Xia</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Navigli</surname>
          </string-name>
          (Eds.),
          <article-title>Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021</article-title>
          ,
          <article-title>Association for Computational Linguistics</article-title>
          , Online,
          <year>2021</year>
          , pp.
          <fpage>944</fpage>
          -
          <lpage>956</lpage>
          . URL: https://aclanthology.org/
          <year>2021</year>
          .findings-acl.
          <volume>82</volume>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <year>2021</year>
          . findings- acl.82.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>H.</given-names>
            <surname>Yuan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Yuan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Yu</surname>
          </string-name>
          ,
          <article-title>Generative biomedical entity linking via knowledge base-guided pre-training and synonyms-aware fine-tuning</article-title>
          , in: M.
          <string-name>
            <surname>Carpuat</surname>
            , M. de Marnefe,
            <given-names>I. V. M.</given-names>
          </string-name>
          <string-name>
            <surname>Ruíz</surname>
          </string-name>
          (Eds.),
          <source>Proceedings of the</source>
          <year>2022</year>
          <article-title>Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</article-title>
          ,
          <string-name>
            <surname>NAACL</surname>
          </string-name>
          <year>2022</year>
          , Seattle, WA,
          <string-name>
            <surname>United</surname>
            <given-names>States</given-names>
          </string-name>
          ,
          <source>July 10-15</source>
          ,
          <year>2022</year>
          , Association for Computational Linguistics,
          <year>2022</year>
          , pp.
          <fpage>4038</fpage>
          -
          <lpage>4048</lpage>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <year>2022</year>
          .naacl- main.296.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Hu</surname>
          </string-name>
          ,
          <article-title>Improving biomedical entity linking with cross-entity interaction</article-title>
          ,
          <source>Proceedings of the AAAI Conference on Artificial Intelligence</source>
          <volume>37</volume>
          (
          <year>2023</year>
          )
          <fpage>13869</fpage>
          -
          <lpage>13877</lpage>
          . URL: https://ojs.aaai.org/index.php/AAAI/article/view/26624. doi:
          <volume>10</volume>
          .1609/aaai.v37i11.
          <fpage>26624</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>J.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Tang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W. X.</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Wei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N. J.</given-names>
            <surname>Yuan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.-R.</given-names>
            <surname>Wen</surname>
          </string-name>
          ,
          <article-title>Few-shot knowledge graph-to-text generation with pretrained language models, in: Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021</article-title>
          ,
          <article-title>Association for Computational Linguistics</article-title>
          , Online,
          <year>2021</year>
          , pp.
          <fpage>1558</fpage>
          -
          <lpage>1568</lpage>
          . URL: https://aclanthology.org/
          <year>2021</year>
          .findings-acl.
          <volume>136</volume>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <year>2021</year>
          .findings- acl.136.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>L.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Petroni</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Josifoski</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Riedel</surname>
          </string-name>
          , L. Zettlemoyer,
          <article-title>Scalable zero-shot entity linking with dense entity retrieval</article-title>
          , in: B.
          <string-name>
            <surname>Webber</surname>
            , T. Cohn,
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>He</surname>
          </string-name>
          , Y. Liu (Eds.),
          <source>Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing, EMNLP</source>
          <year>2020</year>
          , Online,
          <source>November 16-20</source>
          ,
          <year>2020</year>
          , Association for Computational Linguistics,
          <year>2020</year>
          , pp.
          <fpage>6397</fpage>
          -
          <lpage>6407</lpage>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <year>2020</year>
          . emnlp- main.519.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>L.</given-names>
            <surname>Logeswaran</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Toutanova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Devlin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <article-title>Zero-shot entity linking by reading entity descriptions</article-title>
          , in: A.
          <string-name>
            <surname>Korhonen</surname>
            ,
            <given-names>D. R.</given-names>
          </string-name>
          <string-name>
            <surname>Traum</surname>
          </string-name>
          , L. Màrquez (Eds.),
          <source>Proceedings of the 57th Conference of the Association for Computational Linguistics</source>
          ,
          <string-name>
            <surname>ACL</surname>
          </string-name>
          <year>2019</year>
          , Florence, Italy,
          <source>July 28- August 2</source>
          ,
          <year>2019</year>
          , Volume
          <volume>1</volume>
          :
          <string-name>
            <given-names>Long</given-names>
            <surname>Papers</surname>
          </string-name>
          ,
          <source>Association for Computational Linguistics</source>
          ,
          <year>2019</year>
          , pp.
          <fpage>3449</fpage>
          -
          <lpage>3460</lpage>
          . doi:
          <volume>10</volume>
          .18653/v1/p19-
          <fpage>1335</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>T.</given-names>
            <surname>Ayoola</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Tyagi</surname>
          </string-name>
          , J. Fisher,
          <string-name>
            <given-names>C.</given-names>
            <surname>Christodoulopoulos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Pierleoni</surname>
          </string-name>
          ,
          <string-name>
            <surname>Refined:</surname>
          </string-name>
          <article-title>An eficient zeroshot-capable approach to end-to-end entity linking</article-title>
          , in: A.
          <string-name>
            <surname>Loukina</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Gangadharaiah</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          Min (Eds.),
          <source>Proceedings of the 2022 Conference of the North American Chapter of the Association for</source>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>