<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Nested Named Entity Recognition using Multilayer BERT-based Model</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Hasin Rehana</string-name>
          <email>hasin.rehana@und.edu</email>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff4">4</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Benu Bansal</string-name>
          <email>benu.bansal@und.edu</email>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Nur Bengisu Çam</string-name>
          <email>bengisu.cam@std.bogazici.edu.tr</email>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Jie Zheng</string-name>
          <email>jiezhen@med.umich.edu</email>
          <xref ref-type="aff" rid="aff5">5</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Yongqun He</string-name>
          <email>yongqunh@med.umich.edu</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff5">5</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Arzucan Özgür</string-name>
          <email>arzucan.ozgur@bogazici.edu.tr</email>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Junguk Hur</string-name>
          <email>junguk.hur@med.und.edu</email>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Center for Computational Medicine and Bioinformatics, University of Michigan</institution>
          ,
          <addr-line>Ann Arbor, Michigan, 48109</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Department of Biomedical Engineering, University of North Dakota</institution>
          ,
          <addr-line>Grand Forks, North Dakota, 58202</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Department of Biomedical Sciences, School of Medicine and Health Sciences, University of North Dakota</institution>
          ,
          <addr-line>Grand Forks, North Dakota, 58202</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>Department of Computer Engineering, Bogazici University</institution>
          ,
          <addr-line>Istanbul, 34342</addr-line>
          ,
          <country country="TR">Turkey</country>
        </aff>
        <aff id="aff4">
          <label>4</label>
          <institution>School of Electrical Engineering &amp; Computer Science, University of North Dakota</institution>
          ,
          <addr-line>Grand Forks, North Dakota, 58202</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
        <aff id="aff5">
          <label>5</label>
          <institution>Unit for Laboratory Animal Medicine, Department of Microbiology and Immunology, University of Michigan</institution>
          ,
          <addr-line>Ann Arbor, Michigan, 48109</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2024</year>
      </pub-date>
      <abstract>
        <p>In natural language processing, named entity recognition (NER) is a crucial task involving finding and categorizing text entities. The biomedical domain presents substantial hurdles due to the complex structure of the language and the existence of nested entities. This paper introduces an innovative method for Nested NER by utilizing a multilayer bidirectional encoder representation transformer (BERT)-based model, notably employing pretrained PubMedBERT. Our proposed model is designed to manage nested entities' complexities efectively. We combined the robust contextual embeddings from PubMedBERT with a multilayer tagging process. This approach allowed the model to precisely diferentiate between overlapping items, a frequent occurrence in biomedical literature. To assess the efectiveness of our Multilayer NER Model (MultilayerNERModel), we conducted thorough experiments on the BioNNE English Dataset, a dataset for a shared task of BioASQ competition. The findings suggest that employing a multilayer approach enhances the model's ability to identify nested entities, resulting in the thorough detection of entities in biomedical texts. It earned the highest overall performance in English oriented track, with an F1 score of 67.30% and a macro F1 score of 56.36%. These results demonstrate the significant impact of utilizing a multilayer approach in Nested NER tasks, especially in the biomedical domain. The use of UMLS dictionaries, along with the MultilayerNERModel, further enhances the model's performance in biomedical entity recognition.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Named entity recognition (NER)</kwd>
        <kwd>Nested NER</kwd>
        <kwd>Bidirectional encoder representation transformer (BERT)</kwd>
        <kwd>Natural language processing (NLP)</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Identifying and classifying entities, including but not limited to medical terms, names of people,
organizations, and locations, is a critical undertaking in NLP. Nested NER extends this challenge by
requiring the identification of entities embedded within other entities, adding complexity, particularly
in specialized domains such as biomedicine [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. In biomedical text mining, accurate Nested NER
systems are essential for extracting meaningful information from scientific literature, which is crucial
for advancing research and clinical practice [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. The biomedical nested named entity recognition
(BioNNE) task [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] was introduced to address this need as part of the BioASQ Workshop at CLEF 2024
[
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. This task focuses on developing and evaluating NER systems capable of handling nested entities
within biomedical texts. The BioNNE task includes three tracks: English, Russian, and Bilingual. Our
participation was primarily in Track 2 - English-track, though we also applied our approach to the other
tracks. This English track required participants to develop a Nested NER model for English biomedical
scientific abstracts. Participants could train any model architecture on any data provided by organizers
to achieve the best performance, fostering innovation and applying diverse methods.
      </p>
      <p>
        Nested NER has been thoroughly investigated in the domain of NLP while biomedical Nested NER
aims to identify entities such as proteins, genes, diseases, and drugs within biomedical literature [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ].
Traditional methods, such as rule-based approaches and early machine learning models, have been
gradually substituted with advanced techniques that employ deep learning and pretrained language
models. One of the pioneering works in biomedical NER is the introduction of BioBERT [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], a variant of
the BERT model, specifically pretrained on biomedical literature from PubMed [
        <xref ref-type="bibr" rid="ref7 ref8">7, 8</xref>
        ]. BioBERT
demonstrated significant improvements over previous models in various biomedical NER tasks, highlighting
the efectiveness of domain-specific pre-training [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. PubMedBERT was developed as a domain-specific
model following the success of BioBERT [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]. It was trained purely on PubMed abstracts and designed
to enhance the precision and efectiveness of biological NLP activities by utilizing a larger and more
targeted dataset.
      </p>
      <p>
        Nested NER addresses identifying entities embedded within other entities, which is a frequent
occurrence in biomedical texts [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]. Conventional flat NER models lack the necessary capabilities to
handle such intricacies. Various methods have been suggested to address the issue of Nested NER, such
as layered models, span-based models, and sequence-to-sequence models [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ].
      </p>
      <p>FINDING</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>
        MMBERT is a transformer-based model designed to improve the performance of biomedical NER by
integrating multiple models [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ]. It also uses the ERNIE-Health, a Chinese pretrained biomedical
language model. While evaluating MMBERT, they used Chinese biomedical NER datasets. Other than
BERT-based encoder models, GPT-based decoder models are also used for the biomedical domain.
BioGPT [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] is a GPT-2-based biomedical language model that was pretrained on a large amount
GCSs caused a decrease in the serum level of soluble interleukin-2 receptor (sCD25) in both groups.
CHEM
      </p>
      <p>ANATOMY</p>
      <p>CHEM</p>
      <p>CHEM
CHEM
CHEM</p>
      <p>CHEM</p>
      <p>
        CHEM
CHEM
of PubMed articles. BioGPT has been evaluated on six diferent tasks. Another study investigates
the benefits of fine-tuning GPT-3 for biomedical tasks such as NER, relation extraction, and question
answering [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ]. Their experiments on BC5CDR [16], CADEC [17] and ADE [18] showed that their
ifne-tuned GPT-3 models lagged behind the state-of-the-art models.
      </p>
      <p>Some studies have worked on creating datasets designed explicitly for NER tasks involving biomedical
entities. NEREL-BIO [19] is a detailed dataset focusing on nested named entities in the biomedical
domain. It comprises over 700 Russian and 100 English PubMed abstracts, annotated to capture complex
biomedical information through nested entities. The NEREL-BIO has 17 specific biomedical entity
types. The dataset was created as an extension of the general-domain NEREL dataset [20]; therefore,
it is an excellent resource for cross-domain and cross-language benchmarks. Their experiments with
BERT-based and sequence-based models showed that the performance depends on the type of the NER.</p>
      <p>Others introduced a novel bi-encoder framework to improve NER through contrastive learning
[21]. Rather than treating NER as sequence labeling or span classification, the bi-encoder framework
represents the problem as learning vector representations. By mapping candidate text spans and entity
types into the same vector space, the model maximizes the similarity between an entity mentioned and
its type while minimizing the similarity of the non-entity types.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Methodology</title>
      <sec id="sec-3-1">
        <title>3.1. Dataset</title>
        <p>To evaluate the eficacy of our proposed method, we performed experiments using the BioASQ-BioNNE
dataset, which was released in 2024 as a dataset for the shared task challenge of BioASQ [22, 23, 24].
This dataset includes 54 training abstracts, 50 development abstracts, and 500 testing abstracts from
PubMed for the English track. The challenge organizers provided 716 training abstracts, 50 development
abstracts, and 500 testing abstracts for the Russian track. The Bilingual track participants should use
the dataset provided for both English and Russian. The dataset encompasses a total of eight diferent
biomedical named entity classes. For this article, we mainly focused on the English track. Table 1
shows the detailed distribution of diferent entity types across the training and development sets of
the dataset. "DISO" and "ANATOMY" entities are the most frequent term classes, indicating a focus on
anatomical and disorder related information, whereas "DEVICE" entity is the least frequent, suggesting
limited data on medical devices. Since the annotation files for the testing data have not been disclosed
to participants, we do not have the class-wise distribution for the testing set.</p>
        <p>Each abstract in the original dataset is accompanied by a corresponding annotation file. We processed
the dataset by splitting each abstract into sentences and mapping the corresponding annotations to
these sentences. We implemented the BIO-tagging scheme, a well-known method for named entity
recognition encoding. Tokens were encoded as "B-TYPE" for the beginning of an entity, "I-TYPE" for
subsequent tokens of the same entity, and "O" for tokens that do not belong to any entity class. Given
that the provided annotation files indicate up to six nested levels, we applied six levels of BIO-tagging.</p>
        <p>After processing, we shufled the sentences and split the merged dataset into training and validation
sets using an 80:20 ratio. After performing hyperparameter tuning, we combined the training and
validation sets to utilize the entire dataset for training.</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. MultilayerModel</title>
        <p>The core of the method is the MultilayerNERModel, which is our deep learning architecture designed
specifically for Nested NER tasks in biomedical texts. This model is built on the robust foundation
of the pretrained PubMedBERT, a variant of BERT that has been pretrained on a large corpus of the
biomedical literature, making it highly relevant and efective for domain-specific tasks. The overall
architecture of our method is shown in Figure 2. The figure represents a comprehensive workflow for
recognizing and tagging named entities in biomedical texts. It integrates data from the BioASQ-BioNNE
dataset, applies nested BIO-tagging, utilizes a MultilayerNERModel, performs dictionary-based search
using UMLS resources, and concludes with postprocessing and evaluation of the results.</p>
        <p>Abstracts</p>
        <p>Annotations
BioASQ-BioNNE Dataset</p>
        <p>Nested
BIO - Tagging</p>
        <p>MultilayerNER</p>
        <p>Model
Eight UMLS
Dictionaries</p>
        <p>Dictionarybased
Search</p>
        <p>Postprocessed</p>
        <p>Output
Evaluation</p>
        <p>
          Base Model: The base of the model is the pretrained PubMedBERT [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ], which provides
contextualized word embeddings. We also tried state-of-the-art pretrained BERT and BioBERT as an alternative to
PubMedBERT to compare the performance. It can be replaced by any other pretrained model that is
compatible with the dataset domain and language.
        </p>
        <p>Classification Layers: A series of six classification layers were added on top of the base model.
Each layer was designed to output a specific nested level of NER tags, with each linear layer taking
the hidden states from PubMedBERT and mapping them to the required number of labels for that
layer. Although the original number of classes in the BioASQ-BioNNE dataset is eight, to support our
preprocessed BIO-tagged dataset, the total number of output classes for each classification layer is 17.
This includes "B-Class" and "I-Class" for each of the eight original classes, as well as "O" class for the
rest of the tokens in the sentences those do not belong to any entity class.</p>
        <p>To optimize the performance of our method, we conducted hyperparameter tuning and determined
the optimal settings. We utilized the Adam optimizer, known for its eficient handling of sparse gradients
and adaptability to diferent data structures [ 25]. The hyperparameter settings listed in Table 2 were
applied uniformly across all experiments to ensure consistency and facilitate a robust evaluation of our
approach.</p>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. UMLS Dictionaries</title>
        <p>To further enrich the NER process, we leveraged the Unified Medical Language System (UMLS)
Metathesaurus 2024AA for vocabulary expansion [26, 27]. We utilized the MRCONSO.RRF data file within UMLS
to extract relevant concepts and their child concepts based on the UMLS Semantic Group in Table
3, obtained from the NEREL-BIO GitHub repository (https://github.com/nerel-ds/NEREL-BIO/). This
approach allowed us to broaden the model’s ability to recognize entities by incorporating synonyms and
related terms. By integrating these expanded vocabularies into our Nested NER system, we aimed to
enhance the identification and classification of biomedical entities, ultimately improving the robustness
and accuracy of our model.</p>
        <p>We extracted the entities from the UMLS dictionaries that match the concept identifiers of our target
entity types. These dictionaries served as comprehensive references for the various entities we aimed to
identify. Subsequently, we applied these dictionaries to the test data, systematically matching the terms
within the text. By capturing the positions of these terms in the test data, we generated an output file
that listed the identified entities. We merged this output with the results from our MultilayerNERModel
based on BERT, BioBERT, and PubMedBERT. This two-step approach, combining rule-based matching
with advanced machine learning techniques, provided a robust mechanism for entity recognition,
improving the overall quality and reliability of our extracted data.</p>
      </sec>
      <sec id="sec-3-4">
        <title>3.4. Evaluation metrics</title>
        <p>We evaluated the model’s performance using precision, recall, F1 score, and macro F1 score with a
predefined evaluation script at the competition server of BioASQ. These metrics provide a comprehensive
view of the model’s ability to accurately identify and classify named entities in biomedical texts. The
equation for precision, recall, F1 score, and macro F1 score are Eq. (1), (2), (3), and (4), respectively.</p>
        <p>Precision =</p>
        <p>+  
Recall =  
  +</p>
        <p>Precision × Recall</p>
        <p>F1 Score = 2 × Precision + Recall</p>
        <sec id="sec-3-4-1">
          <title>Here, macro F1 score is the average F1 score across the eight entity classes.</title>
        </sec>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Experimental Results</title>
      <p>For this experiment, we employed 6 NVIDIA Tesla V100 GPUs with 32GB of HBM2 VRAM each. The
model training and evaluation were implemented using the PyTorch [28] library. To speed up training,
the DataParallel class was used to leverage multiple GPUs simultaneously.</p>
      <p>The overall distribution of the classes of the dataset after BIO-tagging is illustrated in Figure 3. The
bar chart shows the named entity tag distribution, where "I-FINDING" has the highest frequency, while
"I-INJURY_ POISONING" and "I-DEVICE" have the lowest frequencies. The frequency of "B-DISO" is
higher than "B-ANATOMY". The distribution of "O" is excluded from the figure as they are not the
actual target class of this experiment.</p>
      <p>Figure 4 provides valuable insights into the behavior and performance of diferent levels of nested
entity tags in recognizing various entity types from the dataset. Overall, the frequency of the "B"
(beginning) and "I" (inside) part of "FINDING", "DISO", and "ANATOMY" entities are higher across all
layers, whereas "DEVICE", "PHYS", and "LABPROC" entities show very little or no frequency. The
nested entity level 6 hardly had any entity type other than "O". Again, the frequencies of "O" types
are not included in the bar chart as they are not our original target. These distributions highlight
the specialized capabilities of certain layers in identifying specific entity types, which can inform the
selection of appropriate layers for targeted NER tasks.</p>
      <p>We evaluated the performance of various models, namely baseline BERT, BioBERT, PubMedBERT and
a combination of UMLS knowledge-based dictionary adaptation. The results are summarized in Table
4, presenting the precision, recall, F1 score, and macro F1 score for each model. PubMedBERT-based
MultilayerNERModel achieved the highest overall performance with an F1 score of 67.30% and a macro
F1 score of 56.36%. The score demonstrates its efectiveness in capturing the nuances of biomedical
texts.</p>
      <p>Augmenting the models with UMLS knowledge yielded mixed results. For instance, BioBERT-based
MultilayerNERModel, along with UMLS dictionaries, showed an increase in recall (70.97%) compared to
BioBERT alone (66.28%), suggesting that UMLS integration helps in better entity recognition coverage.</p>
      <sec id="sec-4-1">
        <title>However, this came at the cost of reduced precision (53.58% vs. 64.01%).</title>
        <p>PubMedBERT-based MultilayerNERModel, along with UMLS dictionaries, also demonstrated an
improvement in terms of highest recall (72.55%) over PubMedBERT-based MultilayerNERModel (68.18% ),
but similar to BioBERT, it experienced a drop in precision (55.10% vs. 66.45%). Despite this,
PubMedBERTbased MultilayerNERModel with UMLS achieved a competitive F1 score of 62.63% and a macro F1 score
of 55.46%, indicating that the inclusion of UMLS provides additional benefits in recognizing more
entities, albeit with some trade-ofs in precision.</p>
        <p>We also evaluated our MultilayerNERModel for the Russian and Bilingual Nested NER tracks as
well. For the Russian Nested NER track, we used the pretrained SBERT-Large-NLU-RU [29] as the
base of our MultilayerNERModel, which is a pretrained BERT-based model specifically tailored for the
Russian language. This choice was necessary because BERT, BioBERT, and PubMedBERT are designed
for English text only. Our model achieved precision, recall, and F1 scores of 68.59%, 65.34%, and 66.93%,
respectively, on the Russian BioNNE dataset. However, the macro F1 score was lower (60.07%) than the
F1 score, likely due to potential class imbalance issues.</p>
        <p>For the Bilingual Nested NER track, we employed BERT-Base-Multilingual-uncased [30] as the base
of our MultilayerNERModel, which is tailored to understand and represent 102 diferent languages. The
precision, recall, and F1 scores were 60.27%, 57.5%, and 58.89%, with a macro F1 score of 50.53%. The
performance of our MultilayerNERModel for the Bilingual Nested NER track is lower than the English
and Russian Nested NER tracks. Employing a dictionary-based approach may improve the performance
of our MultilayerNERModel for Russian and Bilingual tracks.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusion</title>
      <p>Our study demonstrates that the domain-specific models, specifically BioBERT and PubMedBERT-based
Nested NER models, significantly outperform the baseline BERT-based Nested NER model in terms of
precision, recall, F1 score, and macro F1 score. This improvement underscores the advantage of using
models pretrained on biomedical literature for Nested NER tasks within this specialized domain.</p>
      <p>Integrating UMLS dictionaries enhances recall, suggesting it helps recognize a broader range of
entities. However, the reduction in precision indicates a need for further optimization to balance the
trade-ofs between precision and recall.</p>
      <p>Overall, PubMedBERT stands out as the most efective pretrained model as the base for our
MultilayerNERModel, with a promising potential for further enhancement through the strategic incorporation
of external knowledge bases like UMLS.</p>
      <p>The implications of this research are substantial in the field of biomedical informatics. An improved
Nested NER system can facilitate more efective information extraction, aiding researchers in
uncovering complex relationships within biomedical literature. This advancement can, in turn, accelerate
the discovery of novel insights and advancements in biomedical science. The performance of our
MultilayerNERModel emphasizes the capability of multilayer BERT-based architectures in advancing
NLP applications in the biomedical domain by addressing the challenges of Nested NER.</p>
    </sec>
    <sec id="sec-6">
      <title>6. Acknowledgments</title>
      <p>The study was supported by the US National Institute of Allergy and Infectious Disease (U24AI171008
to Y.H. and J.H.). GEBIP Award of the Turkish Academy of Sciences (to A.Ö.) is gratefully acknowledged.
Code Availability: https://github.com/hurlab/Nested-NER-BERT
[16] J. Li, Y. Sun, R. J. Johnson, D. Sciaky, C.-H. Wei, R. Leaman, A. P. Davis, C. J. Mattingly, T. C.</p>
      <p>Wiegers, Z. Lu, BioCreative v CDR task corpus: a resource for chemical disease relation extraction,
Database 2016 (2016).
[17] S. Karimi, A. Metke-Jimenez, M. Kemp, C. Wang, Cadec: A corpus of adverse drug event
annotations, Journal of biomedical informatics 55 (2015) 73–81.
[18] C. Hiba, E. H. Nfaoui, C. Loqman, Fine-tuning transformer models for adverse drug event
identification and extraction in biomedical corpora: A comparative study, in: International
Conference on Digital Technologies and Applications, Springer, 2024, pp. 957–966.
[19] N. Loukachevitch, S. Manandhar, E. Baral, I. Rozhkov, P. Braslavski, V. Ivanov, T. Batura, E.
Tutubalina, NEREL-BIO: a dataset of biomedical abstracts annotated with nested named entities,
Bioinformatics 39 (2023) btad161.
[20] N. Loukachevitch, E. Artemova, T. Batura, P. Braslavski, V. Ivanov, S. Manandhar, A. Pugachev,
I. Rozhkov, A. Shelmanov, E. Tutubalina, NEREL: a russian information extraction dataset with
rich annotation for nested entities, relations, and wikidata entity links, Language Resources and
Evaluation (2023) 1–37.
[21] S. Zhang, H. Cheng, J. Gao, H. Poon, Optimizing bi-encoder for named entity recognition via
contrastive learning, arXiv preprint arXiv:2208.14565 (2022).
[22] A. Krithara, J. G. Mork, A. Nentidis, G. Paliouras, The road from manual to automatic semantic
indexing of biomedical literature: a 10 years journey, Frontiers in Research Metrics and Analytics
8 (2023).
[23] A. Krithara, A. Nentidis, K. Bougiatiotis, G. Paliouras, BioASQ-QA: A manually curated corpus for
biomedical question answering, Scientific Data 10 (2023) 170.
[24] G. Tsatsaronis, G. Balikas, P. Malakasiotis, I. Partalas, M. Zschunke, M. R. Alvers, D. Weissenborn,
A. Krithara, S. Petridis, D. Polychronopoulos, An overview of the BIOASQ large-scale biomedical
semantic indexing and question answering competition, BMC bioinformatics 16 (2015) 1–28.
[25] D. P. Kingma, J. Ba, Adam: A method for stochastic optimization, 2017-01-29. URL: http://arxiv.</p>
      <p>org/abs/1412.6980. doi:10.48550/arXiv.1412.6980. arXiv:1412.6980 [cs].
[26] A. K. Chanda, T. Bai, Z. Yang, S. Vucetic, Improving medical term embeddings using UMLS
metathesaurus, BMC Medical Informatics and Decision Making 22 (2022-04-29) 114. URL: https:
//doi.org/10.1186/s12911-022-01850-5. doi:10.1186/s12911-022-01850-5.
[27] V. Nguyen, H. Y. Yip, O. Bodenreider, Biomedical vocabulary alignment at scale
in the UMLS metathesaurus, in: Proceedings of the Web Conference 2021, WWW ’21,
Association for Computing Machinery, 2021-06-03, pp. 2672–2683. URL: https://doi.org/10.1145/
3442381.3450128. doi:10.1145/3442381.3450128.
[28] A. Paszke, S. Gross, F. Massa, A. Lerer, J. Bradbury, G. Chanan, T. Killeen, Z. Lin, N. Gimelshein,
L. Antiga, et al., Pytorch: An imperative style, high-performance deep learning library, Advances
in neural information processing systems 32 (2019).
[29] A. Kabaev, S. Khaustov, N. Gorlova, A. Kalmykov, Bert for russian news clustering, Available at:
https://doi. org/10.28995/2075-7182-2021-20-385-390 (2021).
[30] J. Devlin, M. Chang, K. Lee, K. Toutanova, BERT: pre-training of deep bidirectional transformers
for language understanding, CoRR abs/1810.04805 (2018). URL: http://arxiv.org/abs/1810.04805.
arXiv:1810.04805.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>J.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Yoon</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Kim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Kim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Kim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. H.</given-names>
            <surname>So</surname>
          </string-name>
          ,
          <string-name>
            <surname>J. Kang,</surname>
          </string-name>
          <article-title>BioBERT: a pre-trained biomedical language representation model for biomedical text mining</article-title>
          ,
          <source>Bioinformatics</source>
          <volume>36</volume>
          (
          <year>2020</year>
          )
          <fpage>1234</fpage>
          -
          <lpage>1240</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>S.</given-names>
            <surname>Sivarajkumar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H. A.</given-names>
            <surname>Mohammad</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Oniani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Roberts</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Hersh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>He</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Visweswaran</surname>
          </string-name>
          ,
          <string-name>
            <surname>Y. Wang,</surname>
          </string-name>
          <article-title>Clinical information retrieval: A literature review</article-title>
          ,
          <source>Journal of Healthcare Informatics Research</source>
          (
          <year>2024</year>
          )
          <fpage>1</fpage>
          -
          <lpage>40</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>V.</given-names>
            <surname>Davydova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Loukachevitch</surname>
          </string-name>
          , E. Tutubalina,
          <article-title>Overview of bionne task on biomedical nested named entity recognition at bioasq</article-title>
          <year>2024</year>
          ,
          <year>2024</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>10</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>A.</given-names>
            <surname>Nentidis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Katsimpras</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Krithara</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Lima-López</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Farré-Maduell</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Krallinger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Loukachevitch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Davydova</surname>
          </string-name>
          , E. Tutubalina, G. Paliouras,
          <source>Overview of BioASQ</source>
          <year>2024</year>
          :
          <article-title>The twelfth BioASQ challenge on Large-Scale Biomedical Semantic Indexing and Question Answering</article-title>
          , in: L.
          <string-name>
            <surname>Goeuriot</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Mulhem</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          <string-name>
            <surname>Quénot</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Schwab</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Soulier</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          <string-name>
            <surname>Maria Di Nunzio</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Galuščáková</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>García Seco de Herrera</surname>
          </string-name>
          , G. Faggioli, N. Ferro (Eds.),
          <source>Experimental IR Meets Multilinguality, Multimodality, and Interaction. Proceedings of the Fifteenth International Conference of the CLEF Association (CLEF</source>
          <year>2024</year>
          ),
          <year>2024</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>10</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Bi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. A.</given-names>
            <surname>Dip</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Hajialigol</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Kommu</surname>
          </string-name>
          , H. Liu,
          <string-name>
            <given-names>M.</given-names>
            <surname>Lu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <article-title>AI for biomedicine in the era of large language models</article-title>
          ,
          <source>arXiv preprint arXiv:2403.15673</source>
          (
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>S.</given-names>
            <surname>Masoumi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Amirkhani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Sadeghian</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Shahraz</surname>
          </string-name>
          ,
          <article-title>Natural language processing (NLP) to facilitate abstract review in medical research: the application of BioBERT to exploring the 20-year use of NLP in medical research</article-title>
          ,
          <source>Systematic Reviews</source>
          <volume>13</volume>
          (
          <year>2024</year>
          )
          <fpage>107</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>Q.</given-names>
            <surname>Jin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Kim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. C.</given-names>
            <surname>Comeau</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Yeganova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W. J.</given-names>
            <surname>Wilbur</surname>
          </string-name>
          ,
          <string-name>
            <surname>Z. Lu,</surname>
          </string-name>
          <article-title>MedCPT: Contrastive pre-trained transformers with large-scale PubMed search logs for zero-shot biomedical information retrieval</article-title>
          ,
          <source>Bioinformatics</source>
          <volume>39</volume>
          (
          <year>2023</year>
          )
          <article-title>btad651</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>A.</given-names>
            <surname>Turchin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Masharsky</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Zitnik</surname>
          </string-name>
          ,
          <article-title>Comparison of BERT implementations for natural language processing of narrative medical documents</article-title>
          ,
          <source>Informatics in Medicine Unlocked</source>
          <volume>36</volume>
          (
          <year>2023</year>
          )
          <fpage>101139</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Kim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.-H.</given-names>
            <surname>Kim</surname>
          </string-name>
          , Y.
          <string-name>
            <surname>-M. Kim</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Song</surname>
            ,
            <given-names>H. J.</given-names>
          </string-name>
          <string-name>
            <surname>Joo</surname>
          </string-name>
          ,
          <article-title>Predicting medical specialty from text based on a domain-specific pre-trained BERT</article-title>
          ,
          <source>International Journal of Medical Informatics</source>
          <volume>170</volume>
          (
          <year>2023</year>
          )
          <fpage>104956</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Gu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Tinn</surname>
          </string-name>
          , H. Cheng, M. Lucas,
          <string-name>
            <given-names>N.</given-names>
            <surname>Usuyama</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Naumann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Gao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Poon</surname>
          </string-name>
          ,
          <article-title>Domainspecific language model pretraining for biomedical natural language processing</article-title>
          ,
          <source>ACM Transactions on Computing for Healthcare (HEALTH) 3</source>
          (
          <issue>2021</issue>
          )
          <fpage>1</fpage>
          -
          <lpage>23</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>Q.</given-names>
            <surname>Zheng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Shi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Dong</surname>
          </string-name>
          ,
          <article-title>Exploring interactive and contrastive relations for nested named entity recognition</article-title>
          ,
          <source>IEEE/ACM Transactions on Audio, Speech, and Language Processing</source>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>X.</given-names>
            <surname>Du</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Xing</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Jia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Zan</surname>
          </string-name>
          ,
          <article-title>MRC-based nested medical NER with co-prediction and adaptive pre-training</article-title>
          ,
          <source>arXiv preprint arXiv:2403.15800</source>
          (
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>L.</given-names>
            <surname>Fu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Weng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , H. Xie,
          <string-name>
            <surname>Y. Cao,</surname>
          </string-name>
          <article-title>MMBERT: a unified framework for biomedical named entity recognition</article-title>
          ,
          <source>Medical &amp; Biological Engineering &amp; Computing</source>
          <volume>62</volume>
          (
          <year>2024</year>
          -01-01)
          <fpage>327</fpage>
          -
          <lpage>341</lpage>
          . URL: https://doi.org/10.1007/s11517-023-02934-8. doi:
          <volume>10</volume>
          .1007/s11517-023-02934-8.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>R.</given-names>
            <surname>Luo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Xia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Qin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , H. Poon, T.-Y. Liu,
          <article-title>BioGPT: generative pre-trained transformer for biomedical text generation and mining</article-title>
          ,
          <source>Briefings in bioinformatics 23</source>
          (
          <year>2022</year>
          )
          <article-title>bbac409</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>H.</given-names>
            <surname>Bousselham</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mourhir</surname>
          </string-name>
          ,
          <article-title>Fine-tuning GPT on biomedical NLP tasks: An empirical evaluation</article-title>
          , in: 2024 International Conference on Computer, Electrical &amp; Communication
          <string-name>
            <surname>Engineering</surname>
          </string-name>
          (ICCECE), IEEE,
          <year>2024</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>6</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>