<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Acronym Identification using Transformers and Flair Framework</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>F. Balouchzahi</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>O. Vitman</string-name>
          <email>ovitman2021@cic.ipn.mx</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>H.L. Shashirekha</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>G. Sidorov</string-name>
          <email>sidorov@cic.ipn.mx</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>A. Gelbukh</string-name>
          <email>gelbukh@cic.ipn.mx</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Computer Science, Mangalore University</institution>
          ,
          <addr-line>Mangalore</addr-line>
          ,
          <country country="IN">India</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Instituto Politécnico Nacional (IPN), Centro de Investigación en Computación (CIC)</institution>
          ,
          <addr-line>Mexico City</addr-line>
          ,
          <country country="MX">Mexico</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>The amount of acronyms in texts is growing with the increase in the number of scientific articles and it is not bound only to English texts. The Acronym Extraction (AE) task aims at automatically identifying and extracting the acronyms and their long forms in the given text. To tackle the challenge of AE in diferent languages, this paper describes the participation of the team MUCIC in the AE shared task at the AAAI-22 Workshop on Scientific Document Understanding (SDU@AAAI-22). This shared task aims at identifying and extracting acronyms and their long forms from English, Spanish, French, Danish, Persian, and Vietnamese texts. The proposed methodology consists of data transformation using Spacy and/or other libraries depending on the language and a Flair framework to fine-tune the transformers of the corresponding languages to extract acronyms and their long-forms. For the Spanish language, the proposed methodology secured the second rank and for all other languages, the results obtained are reasonable.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Acronym</kwd>
        <kwd>Expansion</kwd>
        <kwd>Flair</kwd>
        <kwd>BERT</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <sec id="sec-1-1">
        <title>1https://github.com/amirveyseh/AAAI-22-SDU-shared-task-1</title>
        <p>AE</p>
      </sec>
    </sec>
    <sec id="sec-2">
      <title>3. Methodology</title>
      <p>texts in six languages, namely: English, Spanish, French, they adopted several strategies including dynamic
negDanish, Persian, and Vietnamese. ative sample selection, task adaptive pretraining,
adver</p>
      <p>
        The proposed methodology to identify acronyms in sarial training and pseudo-labeling for AD. The
experithe given text contains Data transformation and Model ments conducted won the first place in AD shared task
Fine-Tuning and is based on our previous work [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] that at SDU@AAAI-2021 with F1-score of 0.94.
utilized Flair framework to fine-tune transformers. Our Three models based on Bidirectional Long Short-Term
proposed model obtained promising results for almost Memory (BiLSTM) and Conditional Random Field (CRF)2,
all high-resource languages and the best performance is namely: BiLSTM with CRF Huang et al. [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ], Stacked
achieved for Spanish with a F1-score of 0.90 leading to BiLSTM and CRF Lample et al. [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ], and Bi-LSTM and
second rank in the AE shared task. CRF with convolution and max-pooling Ma et al. [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ]
      </p>
      <p>
        The rest of the paper is organized as follows: Sec- were adopted by Rogers et al. [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ] for AI shared task with
tion 2 describes some of the good performing models Glove embedding for all the models. They also employed
submitted to Acronym Identification (AI) shared task at four transformer models, namely: BERT, BioBERT,
DistilAAAI-21 Workshop on Scientific Document Understand- BERT, and RoBERTa as well for AI shared task. The best
ing (SDU@AAAI-21) followed by the proposed method- performance was obtained using stacked BiLSTM with
ology in Section 3. Experiments and results are discussed CRF with a F1-score of 0.91.
in Section 4 and the paper concludes in Section 5. Despite several models, the complexity of AI/AE
provides scope for further experimentation.
      </p>
    </sec>
    <sec id="sec-3">
      <title>2. Related Work</title>
      <sec id="sec-3-1">
        <title>Researchers have developed several eficient models start</title>
        <p>
          ing from traditional rule-based to advanced DL meth- The proposed methodology is adopted from our
previods for AI, AE and Acronym Disambiguation (AD) tasks. ous work on Automatic Detection of Occupations and
Given an acronym and several possible expansions, AD Profession in Medical Texts using Flair and BERT [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ]
task has to determine the correct expansion for the given applied only on Spanish language texts. With minor
modcontext. AD task is challenging due to the high ambigu- ifications to the existing architecture, the methodology is
ity of acronyms. The organizers of SDU@AAAI-21 have extended for the AE task in six languages text provided
released two large datasets of English scientific papers by the organizers. The workflow of the methodology
conpublished at arXiv for two shared tasks: AI [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ] and AD tains two major parts: Data Transformation and Model
[
          <xref ref-type="bibr" rid="ref13">13</xref>
          ]. The studies and models related to AI, AE and AD Fine-Tuning, which are explained in the following
subare described below: sections:
        </p>
        <p>
          Traditional approaches of sequence labeling, mainly
rule-based or feature-based, are introduced by Schwartz 3.1. Data Transformation
et al. [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ] for AI. Their model builds a dictionary of
local acronyms by utilizing character-match between This phase contains the necessary steps to transform the
acronym letters and corresponding long-forms in the data to a representation that can be used to train and
same sentence to discover the acronym and its long-form. fine-tune the model. The data provided for our previous
        </p>
        <p>
          Zhu et al. [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ] proposed AT-BERT - a Bidirectional En- work [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ] was in Brat standof format 3 and this data
coder Representations from Transformers (BERT)-based was transformed to CONLL IOB4 format as it is easy to
model for AI shared task at SDU@AAAI-21. A Fast Gra- process data in CONLL IOB format rather than in Brat
dient Method (FGM)-based adversarial training strategy format. Brat format consists of a collection of text (.txt)
was incorporated in the fine-tuning of BERT variants, and their corresponding annotation files (.ann).
and an average ensemble mechanism was devised to cap- The datasets for the AI shared task consists of JSON
ture the better representation from multi-BERT variants. files. Each JSON file contains a collection of 4
compoThis proposed model secured first rank in AI shared task nents comprising of text, beginning and ending ofsets
with an average macro F1-score of 0.94. of acronyms and their corresponding long-forms and
        </p>
        <p>
          The model proposed by Egan et al. [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ] uses a trans- an id of that text. A sample JSON file is shown in
Figformer followed by linear projection for AI and finds ure 1. These JSON files are first transformed to Brat
similar examples with embeddings learned from Twin representation as shown in Figure 2 and then the Brat
Networks for AD. With ensemble of diferent transform- representations are transformed to CONLL IOB
repreers, the models obtained F1-scores of 0.93 and 0.91 for AI sentation as described in [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ] and is shown in Figure 3.
and AD shared tasks respectively.
        </p>
        <p>
          Pan et al. [
          <xref ref-type="bibr" rid="ref17">17</xref>
          ] introduces a binary classification model
for AD. Using BERT encoder for input representations,
        </p>
      </sec>
      <sec id="sec-3-2">
        <title>2https://github.com/guillaumegenthial/tf_ner 3https://brat.nlplab.org/standof.html 4https://nlp.lsi.upc.edu/freeling/node/83</title>
        <p>3.2. Model Fine-Tuning</p>
      </sec>
      <sec id="sec-3-3">
        <title>Model Fine-Tuning employs Flair framework to fine-tune</title>
        <p>the pre-trained transformer language model to build a
sequence tagger for the task of AE - a downstream task.</p>
        <p>
          Figure 2: Transformation of data from JSON to Brat format Flair8 is a PyTorch based NLP tool that provides the
facility of utilizing individual or combination of word
embeddings and language models [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ]. Sequence Tagger
module from Flair has BiLSTM backend with CRF layer
The input JSON files of all the languages in the given on top of this model (which is not used in this work).
dataset are first converted to a collection of text (.txt) and Since fine-tuning the transformers is time
consumtheir corresponding annotation files (.ann) according to ing and require significant resources such as RAM and
Brat format based on the provided beginning and ending GPU, models are fine-tuned only for 3 epochs which may
ofsets corresponding to acronyms and their long-forms. probably lead to lower results. As the overall
perforAs the proposed methodology is based on our previous mance of the proposed methodology also depends on the
work, the direct transformation of JSON files to CONLL language model, for each language, the most popular
lanIOB format is avoided. guage model is selected and fine-tuned. The pre-trained
        </p>
        <p>Spacy5 library which provides various tools for pro- transformer language models used for each language
cessing texts in diferent languages is used specifically are presented in Table 1 and the overview of proposed
to extract tokens and sentences from text. However, as methodology is shown in Figure 4.
Spacy does not support low resource languages such as
Persian and Vietnamese, the tools pyvi6 and HAZM7 are
used to extract tokens and sentences from Vietnamese 4. Experiments and Results
and Persian texts respectively.</p>
      </sec>
      <sec id="sec-3-4">
        <title>5https://spacy.io/ 6https://pypi.org/project/pyvi/ 7https://github.com/sobhe/hazm</title>
      </sec>
      <sec id="sec-3-5">
        <title>The primary requirement to promote research in any</title>
        <p>NLP task is the availability of annotated dataset. AE
shared task organizers have provided the participants
with labeled training and development set as well as
unlabelled test set for evaluating the developed models. The</p>
      </sec>
      <sec id="sec-3-6">
        <title>8https://github.com/flairNLP/flair</title>
        <p>
          datasets are provided in six languages, namely: English, The reason for lower results in Persian and Vietnamese
Spanish, Danish, French, Persian, and Vietnamese and could be due to the presence of only acronyms and their
only English language dataset consists of legal and scien- long forms in English (in some cases no long forms also)
tific domains [
          <xref ref-type="bibr" rid="ref22">22</xref>
          ]. Description of the datasets is available and the rest of the text in their native script. As the
transin the GitHub page9 and their statistics are shown in formers used for these languages are monolingual, they
Table 2. It can be observed that the datasets are highly usually do not support other scripts. The proposed model
imbalanced. Further, more number of samples in lan- obtained its best performance in Spanish language and
guages such as Spanish and French may lead to better obtained second rank in the shared task.
performance of the task as compared to less number of Comparison of macro-averaged F1-scores of the top
samples in Vietnamese and Persian languages. models in the shared task for all languages is illustrated
        </p>
        <p>The models submitted to the shared task are evalu- in Figure 5. It can be observed that, as per the
expecated on the blinded test set for predicting the boundaries tations most of models obtained higher performance in
of acronyms and their long-forms based on the macro- English and Spanish languages. The results also prove
averaged scores such as Precision, Recall and F1-score. that as the proposed methodology with only 3 epochs
Participating teams are ranked based on macro-averaged training has shown promising results, experiments could
F1-score and the results obtained by the proposed method be conducted on improving the results by increasing the
for all languages are presented in Table 3. As expected epochs.
the proposed method obtained lower results in Persian
and Vietnamese languages (Spacy does not support these
languages) compared to the results in other languages. 5. Conclusion and Future Work</p>
      </sec>
      <sec id="sec-3-7">
        <title>This paper provides the description of the methodology</title>
        <p>9https://github.com/amirveyseh/AAAI-22-SDU-shared-task-1- and the results obtained by team MUCIC for AE shared
task at SDU@AAAI-22. Data transformation which deals
with diferent data representations is the primary step
in this methodology. The sentences and tokens required
for this step are extracted using Spacy or other libraries
depending on the language. Flair framework used for
ifne-tuning the pre-trained transformer language model
for NER task is extended by building a sequence tagger to
extract acronyms and their long forms. Results obtained
for diferent languages prove that more number of
samples in the training set leads to the higher performances
in identifying the acronyms and their long-forms. The
proposed model obtained its best performance in Spanish
language and obtained second rank in the shared task
and for all other languages, the results obtained are quite
reasonble. As future work we would like to experiment
the combination of embeddings and language models
using Flair frame work as well as other DL methods for
the task of AE in diferent languages.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Acknowledgments</title>
      <sec id="sec-4-1">
        <title>The work was done with partial support from the Mex</title>
        <p>ican Government through the grant A1-S-47854 of the
CONACYT, Mexico, grants 20211784, 20211884, and
20211178 of the Secretaría de Investigación y Posgrado
of the Instituto Politécnico Nacional, Mexico. The
authors thank the CONACYT for the computing resources
brought to them through the Plataforma de Aprendizaje
Profundo para Tecnologías del Lenguaje of the
Laboratorio de Supercómputo of the INAOE, Mexico and
acknowledge the support of Microsoft through the Microsoft Latin
America PhD Award.</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>C. A.</given-names>
            <surname>Mack</surname>
          </string-name>
          ,
          <article-title>How to Write a Good Scientific Paper: Acronyms</article-title>
          , Journal of micro/nanolithography, MEMS, and MOEMS 11 (
          <year>2012</year>
          )
          <fpage>040102</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>A.</given-names>
            <surname>Barnett</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Doubleday</surname>
          </string-name>
          ,
          <article-title>The Growth of Acronyms in the Scientific Literature</article-title>
          ,
          <source>eLife Sciences Publications, Ltd</source>
          <volume>9</volume>
          (
          <year>2020</year>
          )
          <article-title>e60080</article-title>
          . URL: https://doi.org/10. 7554/eLife.60080. doi:
          <volume>10</volume>
          .7554/eLife.60080.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>K.</given-names>
            <surname>Taghva</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Gilbreth</surname>
          </string-name>
          ,
          <source>Finding Acronyms and their Definitions, IJDAR</source>
          <volume>1</volume>
          (
          <year>1999</year>
          )
          <fpage>191</fpage>
          -
          <lpage>198</lpage>
          . doi:
          <volume>10</volume>
          .1007/ s100320050018.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>J.</given-names>
            <surname>Liu</surname>
          </string-name>
          , C. Liu,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Huang</surname>
          </string-name>
          <article-title>, Multi-Granularity Sequence Labeling Model for Acronym Expansion Identification</article-title>
          ,
          <source>Information Sciences 378</source>
          (
          <year>2017</year>
          )
          <fpage>462</fpage>
          -
          <lpage>474</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>K.</given-names>
            <surname>Jacobs</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Itai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Wintner</surname>
          </string-name>
          , Acronyms: Identiifcation, Expansion and Disambiguation,
          <source>Annals of Mathematics and Artificial Intelligence</source>
          <volume>88</volume>
          (
          <year>2020</year>
          )
          <fpage>517</fpage>
          -
          <lpage>532</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>K.</given-names>
            <surname>Kirchhof</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. M.</given-names>
            <surname>Turner</surname>
          </string-name>
          ,
          <article-title>Unsupervised Resolution of Acronyms and Abbreviations in Nursing Notes using Document-level Context Models</article-title>
          ,
          <source>in: Proceedings of the Seventh International Workshop on Health Text Mining and Information Analysis</source>
          ,
          <year>2016</year>
          , pp.
          <fpage>52</fpage>
          -
          <lpage>60</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>J.</given-names>
            <surname>Charbonnier</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Wartena</surname>
          </string-name>
          ,
          <article-title>Using Word Embeddings for Unsupervised Acronym Disambiguation</article-title>
          ,
          <source>in: Proceedings of the 27th International Conference on Computational Linguistics</source>
          ,
          <year>2018</year>
          , pp.
          <fpage>2610</fpage>
          -
          <lpage>2619</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>M.</given-names>
            <surname>Peters</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Neumann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Zettlemoyer</surname>
          </string-name>
          , W.-t. Yih, Dissecting Contextual Word Embeddings:
          <article-title>Architecture and Representation</article-title>
          ,
          <source>in: Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing</source>
          ,
          <year>2018</year>
          , pp.
          <fpage>1499</fpage>
          -
          <lpage>1509</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>J.</given-names>
            <surname>Devlin</surname>
          </string-name>
          , M.-
          <string-name>
            <given-names>W.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Toutanova</surname>
          </string-name>
          , Bert:
          <article-title>Pre-training of Deep Bidirectional Transformers for Language Understanding, in: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</article-title>
          , Volume
          <volume>1</volume>
          (Long and Short Papers),
          <year>2019</year>
          , pp.
          <fpage>4171</fpage>
          -
          <lpage>4186</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>S.</given-names>
            <surname>Y. R. J. F. D. T. H. N. Amir Pouran Ben Veyseh</surname>
          </string-name>
          , Nicole Meister,
          <source>Multilingual Acronym Extraction and Disambiguation Shared Tasks at SDU</source>
          <year>2022</year>
          ,
          <source>in: Proceedings of SDU@AAAI-22</source>
          ,
          <year>2022</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>F.</given-names>
            <surname>Balouchzahi</surname>
          </string-name>
          , G. Sidorov,
          <string-name>
            <given-names>H. L.</given-names>
            <surname>Shashirekha</surname>
          </string-name>
          ,
          <article-title>ADOP FERT-Automatic Detection of Occupations and Profession in Medical Texts using Flair and BERT</article-title>
          ,
          <source>in: Proceedings of the Iberian Languages Evaluation Forum (IberLEF</source>
          <year>2021</year>
          )
          <article-title>co-located with the Conference of the Spanish Society for Natural Language Processing (SEPLN 2021), XXXVII International Conference of the Spanish Society for Natural Language Processing</article-title>
          .,
          <string-name>
            <surname>Málaga</surname>
          </string-name>
          , Spain, September,
          <year>2021</year>
          , volume
          <volume>2943</volume>
          <source>of CEUR Workshop Proceedings, CEUR-WS.org</source>
          ,
          <year>2021</year>
          , pp.
          <fpage>747</fpage>
          -
          <lpage>757</lpage>
          . URL: http: //ceur-ws.
          <source>org/</source>
          Vol-
          <volume>2943</volume>
          /meddoprof_paper2.pdf .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <surname>A. P. B. Veyseh</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Dernoncourt</surname>
            ,
            <given-names>T. H.</given-names>
          </string-name>
          <string-name>
            <surname>Nguyen</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          <string-name>
            <surname>Chang</surname>
            ,
            <given-names>L. A.</given-names>
          </string-name>
          <string-name>
            <surname>Celi</surname>
          </string-name>
          ,
          <article-title>Acronym Identification and Disambiguation Shared Tasks for Scientific Document Understanding</article-title>
          ,
          <source>in: Proceedings of the Workshop on Scientific Document Understanding colocated with 35th AAAI Conference on Artificial Inteligence, SDU@AAAI</source>
          <year>2021</year>
          ,
          <string-name>
            <given-names>Virtual</given-names>
            <surname>Event</surname>
          </string-name>
          ,
          <source>February</source>
          <volume>9</volume>
          ,
          <year>2021</year>
          , volume
          <volume>2831</volume>
          <source>of CEUR Workshop Proceedings, CEUR-WS.org</source>
          ,
          <year>2021</year>
          . URL: http://ceur-ws.
          <source>org/</source>
          Vol-
          <volume>2831</volume>
          /paper33.pdf .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <surname>A. P. B. Veyseh</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Dernoncourt</surname>
            ,
            <given-names>Q. H.</given-names>
          </string-name>
          <string-name>
            <surname>Tran</surname>
            ,
            <given-names>T. H.</given-names>
          </string-name>
          <string-name>
            <surname>Nguyen</surname>
          </string-name>
          ,
          <article-title>What does this Acronym Mean? Introducing a New Dataset for Acronym Identification and Disambiguation</article-title>
          ,
          <source>in: Proceedings of COLING</source>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>A.</given-names>
            <surname>Schwartz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Hearst</surname>
          </string-name>
          ,
          <article-title>A Simple Algorithm for Identifying Abbreviation Definitions in Biomedical Text</article-title>
          ,
          <source>Pacific Symposium on Biocomputing. Pacific Symposium on Biocomputing</source>
          <volume>4</volume>
          (
          <year>2003</year>
          )
          <fpage>451</fpage>
          -
          <lpage>62</lpage>
          . doi:
          <volume>10</volume>
          .1142/9789812776303_
          <fpage>0042</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>D.</given-names>
            <surname>Zhu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Zhong</surname>
          </string-name>
          , G. Zeng,
          <string-name>
            <given-names>W.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Tang</surname>
          </string-name>
          , AT-BERT:
          <article-title>Adversarial Training BERT for Acronym Identification Winning Solution for SDU@AAAI-21</article-title>
          , CEUR Workshop Proceedings (
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>N.</given-names>
            <surname>Egan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Bohannon</surname>
          </string-name>
          ,
          <article-title>Primer AI's Systems for Acronym Identification and Disambiguation</article-title>
          ,
          <source>in: Proceedings of the Workshop on Scientific Document Understanding co-located with 35th AAAI Conference on Artificial Inteligence, SDU@AAAI</source>
          <year>2021</year>
          ,
          <string-name>
            <given-names>Virtual</given-names>
            <surname>Event</surname>
          </string-name>
          ,
          <source>February</source>
          <volume>9</volume>
          ,
          <year>2021</year>
          , volume
          <volume>2831</volume>
          <source>of CEUR Workshop Proceedings, CEUR-WS.org</source>
          ,
          <year>2021</year>
          . URL: http://ceur-ws.
          <source>org/</source>
          Vol-
          <volume>2831</volume>
          /paper30.pdf .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>C.</given-names>
            <surname>Pan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Song</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Luo</surname>
          </string-name>
          ,
          <article-title>BERT-based Acronym Disambiguation with Multiple Training Strategies</article-title>
          ,
          <source>in: Proceedings of the Workshop on Scientific Document Understanding co-located with 35th AAAI Conference on Artificial Inteligence, SDU@AAAI</source>
          <year>2021</year>
          ,
          <string-name>
            <given-names>Virtual</given-names>
            <surname>Event</surname>
          </string-name>
          ,
          <source>February</source>
          <volume>9</volume>
          ,
          <year>2021</year>
          , volume
          <volume>2831</volume>
          <source>of CEUR Workshop Proceedings</source>
          , CEURWS.org,
          <year>2021</year>
          . URL: http://ceur-ws.
          <source>org/</source>
          Vol-
          <volume>2831</volume>
          / paper25.pdf .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Yu</surname>
          </string-name>
          ,
          <article-title>Bidirectional lstm-crf models for sequence tagging</article-title>
          ,
          <source>arXiv preprint arXiv:1508</source>
          .
          <year>01991</year>
          (
          <year>2015</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>G.</given-names>
            <surname>Lample</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ballesteros</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Subramanian</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Kawakami</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Dyer</surname>
          </string-name>
          ,
          <article-title>Neural Architectures for Named Entity Recognition</article-title>
          ,
          <source>in: Proceedings of NAACL-HLT</source>
          ,
          <year>2016</year>
          , pp.
          <fpage>260</fpage>
          -
          <lpage>270</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>X.</given-names>
            <surname>Ma</surname>
          </string-name>
          , E. Hovy,
          <article-title>End-to-end Sequence Labeling via Bi-directional LSTM-CNNs-CRF, in: Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics</article-title>
          (Volume
          <volume>1</volume>
          :
          <string-name>
            <surname>Long</surname>
            <given-names>Papers)</given-names>
          </string-name>
          ,
          <year>2016</year>
          , pp.
          <fpage>1064</fpage>
          -
          <lpage>1074</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>W.</given-names>
            <surname>Rogers</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. R.</given-names>
            <surname>Rae</surname>
          </string-name>
          ,
          <string-name>
            <surname>D.</surname>
          </string-name>
          Demner-Fushman,
          <article-title>AI-NLM exploration of the Acronym Identification Shared Task at SDU@ AAAI-21</article-title>
          .,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>S.</given-names>
            <surname>Y. R. J. F. D. T. H. N. Amir Pouran Ben Veyseh</surname>
          </string-name>
          , Nicole Meister,
          <article-title>MACRONYM: A LargeScale Dataset for Multilingual and Multi-Domain Acronym Extraction</article-title>
          , in: arXiv,
          <year>2022</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>