<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Information Extraction from Spanish Radiology Reports using multilingual BERT</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Oswaldo Solarte-Pabón</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Orlando Montenegro</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Alberto Blazquez-Herranz</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Hadi Saputro</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Alejandro Rodriguez-González</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ernestina Menasalvas</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Centro de Tecnología Biomédica, Universidad Politécnica de Madrid</institution>
          ,
          <country country="ES">Spain</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Universidad del Valle</institution>
          ,
          <addr-line>Cali</addr-line>
          ,
          <country country="CO">Colombia</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>This paper describes our team's participation in Task 1 of the Conference and Labs of the Evaluation Forum (CLEF eHealth 2021). The Task 1 challenge targets Named Entity Recognition (NER) from radiology reports written in Spanish. Our approach addresses this challenge as a sequence labeling task and is based on multilingual BERT with a classification layer on top. Three BERT-based models were trained to support overlapping entities extraction: the first model predicts the first specific label annotated in the corpus; the second predicts the second label for tokens that have two diferent annotations; and the third is used for tokens annotated with a third label in the corpus. Our approach obtained 78.47% and 73.27% for a Lenient and the exact F1 score, respectively.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Information Extraction</kwd>
        <kwd>Named Entity Recognition (NER)</kwd>
        <kwd>Multilingual BERT</kwd>
        <kwd>Radiology Reports</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Radiology reports are one of the most important sources of clinical imaging information. They
document critical information about the patient’s health and the radiologist’s interpretation of
medical findings [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. Extracted information from Radiology reports can be used to support
clinical research, quality improvement, and evidence-based medicine [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. However, the information
in radiology reports is presented in free text format, which makes the task of structuring the
data particularly challenging [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. Extracting this information manually is not a viable task as it
is costly and time-consuming [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ].
      </p>
      <p>
        Several studies have been proposed to extract information from radiology reports [
        <xref ref-type="bibr" rid="ref5 ref6 ref7">5, 6, 7</xref>
        ].
Most of these proposals use machine learning methods, mainly based on the Conditional Random
Fields (CRF) algorithm [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], in which entity extraction is considered as a sequence labeling task.
Recently, deep learning approaches have been shown to improve performance at processing
natural language texts [
        <xref ref-type="bibr" rid="ref10 ref11 ref9">9, 10, 11</xref>
        ]. Previous studies have shown significant progress in extracting
information from clinical reports, but most eforts have focused only on the English language.
However, the Task 1 challenge of CLEF 2021 targets Named Entity Recognition (NER) and
Classification from radiology reports written in Spanish, and specifically ultrasounds reports.
[
        <xref ref-type="bibr" rid="ref12 ref13">12, 13</xref>
        ].
      </p>
      <p>
        In this paper, we describe the approach we have followed to deal with Task 1 of the Conference
and Labs of the Evaluation Forum (CLEF 2021). Our approach takes advantage of the transfer
learning technique, which aims to learn new and specific tasks by transferring knowledge from
pre-trained models [
        <xref ref-type="bibr" rid="ref14 ref15 ref16">14, 15, 16</xref>
        ]. Specifically, we use contextualized pre-trained embeddings
from multilingual BERT [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ] with a classification layer to perform NER from radiology reports
written in Spanish. Three classifiers were trained to predict overlapping entities: i) one to
predict the first specific label in the corpus; ii) another classifier to predict the second label for
those tokens that have two diferent annotations; iii) and the third is used for those tokens that
have a third annotated label.
      </p>
      <p>The rest of the paper has been organized as follows: Section 2 reviews the most recent works
on extracting information from radiology reports. Section 3 describes the components of the
proposed approach. Section 4 presents the results of the experiments, and Section 5 presents
the main conclusions and outlook for future work.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related works</title>
      <p>
        Computational analysis of radiology reports has recently gained greater attention. Most of
the research focuses on identifying specific medical conditions contained in a given report,
and they generally deal with the task of classification. Another generic area of research is
developing information extraction models, which try to extract specific information such as
medical recommendations from radiology reports [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ].
      </p>
      <p>
        A system for identifying named entities from the radiology reports was proposed by
Hassanpour [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. The objective of this model was to identify the specific named entities from the
radiology reports based on their information extraction model. They used a CRF-based model
with several auxiliary features, including POS tags and Radlex Lexicons [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ] to identify entities
of vfie diferent classes: Anatomy, Anatomy Modifier, Observation, Modifier, and Uncertainty.
In [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ], an approach was proposed to extract useful information in abdominopelvic radiology
reports. This approach combines Glove word embeddings with a Recurrent Neural Network [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ]
to extract named entities from radiology reports. This approach shows the feasibility of neural
nets for extracting information from a small corpus of 120 abdominopelvic cross-sectional
imaging reports.
      </p>
      <p>
        Most of the proposals mentioned above have been focused on the English language. According
to [
        <xref ref-type="bibr" rid="ref22">22</xref>
        ], information extraction in the medical domain also represents its own challenges in other
languages. In the case of the Spanish language, [
        <xref ref-type="bibr" rid="ref23">23</xref>
        ] proposed a manually annotated corpus
using text data from radiology reports. This corpus aims to support NER in clinical text written
in Spanish.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Data and Methods</title>
      <p>The CLEF eHealth 2021 Task 1 (Multilingual Information Extraction) focuses on NER from
radiology reports written in Spanish, and specifically from ultrasounds reports. Our approach
addresses this challenge as a sequence labeling task, in which each token in a sentence is
classified using the BIO tagging format. In this format, each token is labeled with B (at the
beginning of the entity), I (inside the entity), or O (Outside the entity).
3.1. Data
The CLEF eHealth 2021 Task 1 organizers provided an annotated corpus in BRAT format [24].
This corpus consists of ultrasonography reports provided by a pediatric hospital in Argentina.
Reports were annotated by clinical experts and then revised by linguists. The corpus contains
three sets: i) Training set (175 reports), ii) Development set (92 reports), and iii) Test set (207
reports). Figure 1 shows an example of annotations provided in the corpus for Task 1.</p>
      <p>The corpus provided was annotated using ten diferent labels: "Abbreviation,
Anatomical_Entity, ConditionalTemporal, Degree, Finding, Location, Measure, Negation, Type_of_measure,
and Uncertainty". In this corpus, annotations can be represented by a word or a sequence of
words (not necessarily continuous). Moreover, a same token can be annotated with diferent
labels, this implies that overlapping entities can be presented. From Figure 1, it can be seen that
extracting information from radiology reports written in Spanish has several challenges:
• Overlapping entities: some tokens may have more than one annotated label. For
instance, the token "FID" is annotated with the label "Anatomical_Entity" and, with the
label "Abbreviation". Moreover, overlapping entities can also mean that a token can
belong to more than one entity type. For instance, in the first sentence ( see Figure 1), the
token "cm" is embedded into the labels "Finding" and "Measure". It causes that this token
contains three annotated labels: "Finding, Measure, and Abbreviation".
• Lengthy entities: these entities are composed of many tokens, which is common in the
corpus. One can see that annotations with the label "Finding" in Figure 1 are composed
of many tokens.
• Discontinuities: these occur when an entity contains annotated tokens that are not
continuous in the text document. This is the case of annotations with the label "Measure"
in Figure 1.
• Polysemy: this is also a common linguistic phenomenon in the Task 1 corpus provided.</p>
      <p>It frequently occurs between labels such as "Location" and "Anatomical Entity."</p>
      <sec id="sec-3-1">
        <title>3.2. Methods</title>
        <p>Our approach consists of three steps: Text pre-processing, Entity Extraction, and Post-Processing
(See Figure 2). In the first step, files in BRAT format are converted into BIO format. In the
second, NER models are trained to perform entity extraction, and in the third, predicted entities
from NER models are assembled in BRAT format again as output. Details of each step are
described as follows.</p>
        <sec id="sec-3-1-1">
          <title>3.2.1. Text Pre-processing</title>
          <p>In this step, each file in the corpus is divided into a set of sentences, and each sentence is
tokenized. Next, the sentences are converted into a BIO file format. As the corpus contains
annotated tokens of up to three diferent labels; the generated BIO files also have three columns
that describe the annotations for each token. Table 1 shows an example of the generated BIO
ifle for a specific text sentence taken from the corpus. The column "Label 1" represents the first
annotated label in the corpus for a specific token. The column "Label 2" is used to represent
a second annotated label in the corpus. Those tokens that do not have a second annotation
are assigned the "O" (Outside) tag by default. Lastly, the column "Label 3" represents a third
annotated label in the corpus. For instance, the token "cm" contains three diferent annotated
labels in the corpus; therefore, this token has three diferent BIO labels (see Table 1). The
columns "Label 2" and "Label 3" are used to represent overlapped entities for the tokens that
have more than one annotated label.</p>
        </sec>
        <sec id="sec-3-1-2">
          <title>3.2.2. Entity Extraction</title>
          <p>The purpose of this step is to perform NER from radiology reports using a transformers-based
model[25] with a classification layer on top of it. Specifically, we use the
bert-base-multilingualcased model, which has been pre-trained on data in 104 languages. The input for the BERT
model is a sequence of tokenized and encoded tokens of a sentence extracted previously from
the BIO files created.</p>
          <p>Figure 3 shows the entity extraction process in which three BERT-based models (1, 2, 3)
were trained to support overlapping entities extraction. Each NER model is trained using a
diferent input data column ( 1, 2, 3). The first model is trained with annotations obtained
from the "Label 1" column in the BIO file (see Table 1), the second is trained with annotations
from the "Label 2" column, and the third is trained using annotations from the "Label 3" column.
Each NER model  operates independently and predicts a diferent set of entities ( 1, 2, 3),
depending on the annotations with which it was trained.</p>
          <p>As can be seen in Table 1, there are three data columns containing diferent annotated labels
for a given token, and we trained each NER model using a diferent input data column. Therefore,
each NER model  predicts diferent entities. On the one hand, the first model is trained to
predict all ten labels annotated in the dataset. On the other hand, the second and third models
are used to predict overlapped entities learned from the tokens that have a second and a third
annotated label respectively. Although these models work independently, the fact that they
predict diferent labels helps to solve the task of predicting overlapped labels.</p>
          <p>The BERT-based models are specialized for NER, with a fully connected layer on top of
the hidden states of each token. These models are fine-tuned on the corpus provided. The
training and development sets of the corpus were mixed, and then we trained the models using
a cross-validation strategy with a value of  = 10. The fine-tuning was performed with a
sequence length of 256 tokens, a batch size of 32, and the number of epochs is 10. The original
BERT tokenizer model drove the tokenization process.</p>
        </sec>
        <sec id="sec-3-1-3">
          <title>3.2.3. Post-processing</title>
          <p>The goal in this step is to take as input the predicted entities (1, 2, 3) from the
BERTbased NER models (Figure 3), and to create files in BRAT format with their annotations and
corresponding ofsets.</p>
          <p>Algorithm 1 shows the procedure used to create annotations in BRAT format. This procedure
receives a sentence text, the text file to which the sentence belongs, and the entities extracted
from the NER models. The NER models work at the sentence level. However, to create a file in
BRAT format, it is important to have the original text file in order to obtain the annotations’
ofsets.</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Results</title>
      <p>In the oficial test phase, 207 files with unannotated radiology reports were available for
evaluating the models. Table 2 shows the oficial performance of our proposal for Lenient and the exact
F1 score. We obtained 78.47% and 73.27% for a Lenient and the exact F1 score, respectively.
The proposed approach shows promising results and suggests the feasibility of multilingual
BERT to perform NER from radiology reports written in Spanish.</p>
      <p>Taking into account the performance for each label, Table 2 shows that those with the best
performance were: "Abbreviation, Anatomical_Entity, Negation, and Type_of_measure". These
labels obtained a performance above 80%. We found that the competitive performance of these
labels can be due to:
• These labels do not commonly contain overlapping entities which increases the
performance of entity extraction in comparison with those labels containing overlapping
entities.
• These labels contain a high number of annotations in the corpus in comparison with
other labels. For instance, the label "Anatomical_Entity" has more than 1700 annotations.
This fact suggests that having a corpus with more annotated data is useful to improve
NER performance using deep learning methods such as multilingual BERT.
• In the case of the negation label, note that the proposed approach shows a very good
performance. The "negation" label has a high number of annotations in comparison with
other labels. We found more than 750 annotations with this label.</p>
      <p>In contrast, the lowest-performing labels were "Finding, Uncertainty, ConditionalTemporal,
Degree, Location and Uncertainty", with a performance below 70%. We identified several reasons
for this behavior:
• In the case of the label "Finding", it is used to represent lengthy entities that contain a long
number of tokens. It is clear that the extraction of lengthy entities afects the performance
of the proposed NER models. In this case, the Recall measure is higher than Precision
measure (see Table 2). This fact can be explained by two reasons: The first reason is
because this label frequently contains annotations with a high number of tokens, which
makes extracting all annotated tokens more dificult than in those entities that only have
one annotated token. The second is because the "Finding" label contains annotations with
discontinuous tokens which also reduces the Precision measure.
• The high variability of words annotated with the label "Uncertainty" together with the
low number of annotations, when compared with the "Negation" label, can be the cause
of the low performance in predicting the "Uncertainty" label.
• The low performance of the "ConditionalTemporal" and "Degree" labels is caused by
the low number of annotations in the corpus compared to other labels. We found 75
annotations with the "Degree" label and only 25 with the "ConditionalTemporal" label.
This number of annotations are much lower in comparison with those labels which
obtained better results.</p>
      <sec id="sec-4-1">
        <title>4.1. Error Analysis</title>
        <p>We identified three more common sources of error which could afect the performance of our
approach. These error sources are related to: i) annotations with discontinuous tokens; ii)
the high number of annotations with the label "O" (Outside) in the BIO file; and iii) the high
variability of annotated tokens in those labels with a low number of annotations.
• The first source of error identified is when an annotation contains discontinuous tokens.</p>
        <p>In these cases, predicting the correct annotated label for each token, and obtaining the
correct ofsets, can be dificult. For instance, in the text "VIA BILIAR intra y extrahepatica",
there are two annotations with the label "Anatomical_Entity" in the provided dataset:
"VIA BILIAR extrahepatica" and "VIA BILIAR intra hepatica". However, these two
annotations contain discontinuous tokens and predicting labels for each token fails because
the automatic text tokenizer being used produces diferent tokens to those annotated in
the input text. In this case, our approach takes the string "extrahepatica" as one token,
but in the annotated corpus, it is used as two sub-words: "extra" and "hepatica". A similar
problem occurs when processing the text "Quiste en epididimo izq de: 3,6 x 1,4 x 1,9mm",
where there are three discontinuous annotations with the label "Measure": "3,6 mm", "1,4
mm" and "1,9 mm". In the above examples, the main issue was calculating the exact token
ofsets which match the original text in the dataset.
• The second source of error is the high number of annotated tokens with the label "O"
(Outside) in the BIO file. This problem mainly afects the third NER model ( 3), which
has been trained with annotations from the column "Label 3" in the BIO file (see Table
1). This column is used to represent a third annotation for those tokens which have
overlapped entities. In this case, the number of annotations with the label "O" is much
higher than other labels. This fact causes a reduction in the performance of the third NER
model in comparison with the first and the second model.
• The third source of error is caused by the low number of annotations on certain labels
combined by the high variability of annotated tokens in these labels. This occurs in
labels such as "Uncertainty", "Degree", and "ConditionalTemporal". In these cases, the
Precision measure is higher than the Recall measure. This fact can be explained by the
high variability of tokens used to annotate these labels, which increases the dificulty
of predicting all possible tokens in the extraction process. For instance, comparing the
behavior of the "Uncertainty" label with the "Negation" label, one can see that "Negation"
has less variability in the annotated corpus. In the case of the "Negation" label, tokens
like "no", "sin", "no se observa" are frequently used. On the other hand, the number of
words expressing "Uncertainty" are more disperse which causes a reduction in the Recall
measure. The fact that the "Uncertainty" label has less annotations and more variability
in annotated tokens than the "Negation" label, could be the reason why the "Uncertainty"
label does not perform as well as the "Negation" label.</p>
        <p>Finally, in the proposed approach, the biggest challenge was being able to predict labels that
have overlapping entities. For this purpose, a three-parallel NER architecture was chosen as
explained in Section 3.2.2, which showed promising results. However, before reaching this
solution we tested two other configurations. The first solution was to combine entities with one
annotation with all overlapping entities into a single column in the BIO file. This single column
represented all the annotations in the corpus. However, the main disadvantage of this solution
was the large number of labels that need to be predicted. The second solution was to create
a diferent annotation column for each annotated label. In this case, the BIO file would need
ten columns with annotations, one column for each label in the corpus. The main drawback
of this solution is that for each column of the BIO file, the number of annotations with the
label "O" (Outside), was greater than the label we wanted to predict. The performance of this
solution was much lower than our current proposed approach, so we rejected that solution.
After carrying out those experiments, the approach that obtained the best performance was
to create a combination of three NER models, as we described in section 3.2.2. This proposed
solution allows extracting labels that have up to three overlapping entities, just as the annotated
entities are found in the corpus provided.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusions</title>
      <p>The proposed approach to perform NER from radiology reports has shown promising results,
suggesting that the multilingual BERT is feasible to extract information from clinical resources
written in Spanish. The main advantage of the proposed approach is that it can automatically
extract features from text data, avoiding the time-consuming feature engineering process.</p>
      <p>Extracting information from radiology reports written in Spanish is a complex task. It poses
several linguistic challenges, such as overlapping entities, lengthy entities, and discontinuities.
We deal with these challenges by training three BERT-based models. Although our approach
has shown promising results, some challenges still have to be addressed, such as extracting
lengthy entities, improving the performance to extract labels with a low number of annotations,
and identifying uncertainty in radiology reports.
[24] P. Stenetorp, S. Pyysalo, G. Topić, T. Ohta, S. Ananiadou, J. Tsujii, brat: a web-based tool
for NLP-assisted text annotation, in: Proceedings of the Demonstrations Session at EACL
2012, Association for Computational Linguistics, Avignon, France, 2012, pp. 102–107.
[25] H. Adel, H. Schütze, Exploring diferent dimensions of attention for uncertainty detection,
15th Conference of the European Chapter of the Association for Computational Linguistics,
EACL 2017 - Proceedings of Conference 1 (2017) 22–34. doi:10.18653/v1/e17-1003.
arXiv:1612.06549.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>S.</given-names>
            <surname>Hassanpour</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. P.</given-names>
            <surname>Langlotz</surname>
          </string-name>
          ,
          <article-title>Information extraction from multi-institutional radiology reports</article-title>
          ,
          <source>Artificial Intelligence in Medicine</source>
          <volume>66</volume>
          (
          <year>2016</year>
          )
          <fpage>29</fpage>
          -
          <lpage>39</lpage>
          . URL: http://dx.doi.org/10.1016/ j.artmed.
          <year>2015</year>
          .
          <volume>09</volume>
          .007. doi:
          <volume>10</volume>
          .1016/j.artmed.
          <year>2015</year>
          .
          <volume>09</volume>
          .007.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>W. W.</given-names>
            <surname>Yim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Yetisgen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W. P.</given-names>
            <surname>Harris</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W. K.</given-names>
            <surname>Sharon</surname>
          </string-name>
          ,
          <source>Natural Language Processing in Oncology Review, JAMA Oncology 2</source>
          (
          <year>2016</year>
          )
          <fpage>797</fpage>
          -
          <lpage>804</lpage>
          . doi:
          <volume>10</volume>
          .1001/jamaoncol.
          <year>2016</year>
          .
          <volume>0213</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>O.</given-names>
            <surname>Solarte-Pabon</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Torrente</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Rodriguez-Gonzalez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Provencio</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Menasalvas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Tunas</surname>
          </string-name>
          ,
          <article-title>Lung cancer diagnosis extraction from clinical notes written in spanish</article-title>
          ,
          <source>Proceedings - IEEE Symposium on Computer-Based Medical Systems 2020-July</source>
          (
          <year>2020</year>
          )
          <fpage>492</fpage>
          -
          <lpage>497</lpage>
          . doi:
          <volume>10</volume>
          .1109/CBMS49503.
          <year>2020</year>
          .
          <volume>00099</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>O. Solarte</given-names>
            <surname>Pabón</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Torrente</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Provencio</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Rodríguez-Gonzalez</surname>
          </string-name>
          , E. Menasalvas,
          <article-title>Integrating Speculation Detection and Deep Learning to Extract Lung Cancer Diagnosis from Clinical Notes</article-title>
          ,
          <source>Applied Sciences</source>
          <volume>11</volume>
          (
          <year>2021</year>
          ). URL: https://www.mdpi.com/2076-3417/ 11/2/865. doi:
          <volume>10</volume>
          .3390/app11020865.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>S.</given-names>
            <surname>Hassanpour</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. P.</given-names>
            <surname>Langlotz</surname>
          </string-name>
          ,
          <article-title>Information extraction from multi-institutional radiology reports</article-title>
          ,
          <source>Artificial Intelligence in Medicine</source>
          <volume>66</volume>
          (
          <year>2016</year>
          )
          <fpage>29</fpage>
          -
          <lpage>39</lpage>
          . URL: http://dx.doi.org/10.1016/ j.artmed.
          <year>2015</year>
          .
          <volume>09</volume>
          .007. doi:
          <volume>10</volume>
          .1016/j.artmed.
          <year>2015</year>
          .
          <volume>09</volume>
          .007.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>N.</given-names>
            <surname>Nandhakumar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Sherkat</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E. E.</given-names>
            <surname>Milios</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Gu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Butler</surname>
          </string-name>
          ,
          <article-title>Clinically significant information extraction from radiology reports</article-title>
          ,
          <source>DocEng 2017 - Proceedings of the 2017 ACM Symposium on Document Engineering</source>
          (
          <year>2017</year>
          )
          <fpage>153</fpage>
          -
          <lpage>162</lpage>
          . doi:
          <volume>10</volume>
          .1145/3103010.3103023.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Steinkamp</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Chambers</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Lalevic</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H. M.</given-names>
            <surname>Zafar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T. S.</given-names>
            <surname>Cook</surname>
          </string-name>
          ,
          <article-title>Toward Complete Structured Information Extraction from Radiology Reports Using Machine Learning</article-title>
          ,
          <source>Journal of Digital Imaging</source>
          <volume>32</volume>
          (
          <year>2019</year>
          )
          <fpage>554</fpage>
          -
          <lpage>564</lpage>
          . doi:
          <volume>10</volume>
          .1007/s10278-019-00234-y.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>J. D.</given-names>
            <surname>Laferty</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>McCallum</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F. C. N.</given-names>
            <surname>Pereira</surname>
          </string-name>
          ,
          <article-title>Conditional random fields: Probabilistic models for segmenting and labeling sequence data</article-title>
          ,
          <source>in: Proceedings of the Eighteenth International Conference on Machine Learning</source>
          , ICML '
          <fpage>01</fpage>
          , Morgan Kaufmann Publishers Inc., Williamstown, MA, USA,
          <year>2001</year>
          , pp.
          <fpage>282</fpage>
          -
          <lpage>289</lpage>
          . URL: http://dl.acm.org/citation.cfm?id=
          <volume>645530</volume>
          .
          <fpage>655813</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>G.</given-names>
            <surname>Lample</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ballesteros</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Subramanian</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Kawakami</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Dyer</surname>
          </string-name>
          ,
          <article-title>Neural architectures for named entity recognition, in: Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Association for Computational Linguistics</article-title>
          , San Diego, California,
          <year>2016</year>
          , pp.
          <fpage>260</fpage>
          -
          <lpage>270</lpage>
          . URL: https://www.aclweb.org/anthology/N16-1030. doi:
          <volume>10</volume>
          .18653/v1/
          <fpage>N16</fpage>
          -1030.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>L.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Luo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Wampfler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Yang</surname>
          </string-name>
          , H. Liu,
          <article-title>Natural language processing for populating lung cancer clinical research data</article-title>
          ,
          <source>BMC Medical Informatics and Decision Making</source>
          <volume>19</volume>
          (
          <year>2019</year>
          )
          <fpage>1</fpage>
          -
          <lpage>10</lpage>
          . URL: http://dx.doi.org/10.1186/s12911-019-0931-8. doi:
          <volume>10</volume>
          .1186/ s12911-019-0931-8.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>J.</given-names>
            <surname>Devlin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. W.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Toutanova</surname>
          </string-name>
          , BERT:
          <article-title>Pre-training of deep bidirectional transformers for language understanding</article-title>
          ,
          <source>NAACL HLT</source>
          <year>2019</year>
          <article-title>- 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies -</article-title>
          <source>Proceedings of the Conference</source>
          <volume>1</volume>
          (
          <year>2019</year>
          )
          <fpage>4171</fpage>
          -
          <lpage>4186</lpage>
          . arXiv:
          <year>1810</year>
          .04805.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>H.</given-names>
            <surname>Suominen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Goeuriot</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Kelly</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. A.</given-names>
            <surname>Alemany</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Bassani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Brew-Sam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Cotik</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Filippo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>González-Sáez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Luque</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Mulhem</surname>
          </string-name>
          , G. Pasi,
          <string-name>
            <given-names>R.</given-names>
            <surname>Roller</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Seneviratne</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Upadhyay</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Vivaldi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Viviani</surname>
          </string-name>
          ,
          <string-name>
            <surname>C.</surname>
          </string-name>
          <article-title>Xu, Overview of the clef ehealth evaluation lab 2021, in: Overview of the CLEF eHealth evaluation lab 2021</article-title>
          ,
          <source>in: CLEF 2021 - 12th Conference and Labs of the Evaluation Forum, Lecture Notes in Computer Science (LNCS)</source>
          , Springer,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>V.</given-names>
            <surname>Cotik</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. A.</given-names>
            <surname>Alemany</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Filippo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Luque</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Roller</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Vivaldi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Ayach</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Carranza</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. D.</given-names>
            <surname>Francesca</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Dellanzo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. F.</given-names>
            <surname>Urquiza</surname>
          </string-name>
          ,
          <article-title>Overview of the clef ehealth evaluation lab 2021, in: Overview of CLEF eHealth Task 1 - SpRadIE: A challenge on information extraction from Spanish Radiology Reports, in: CLEF 2021 Evaluation Labs</article-title>
          and Workshop: Online Working Notes, CEUR-WS, Springer,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>F.</given-names>
            <surname>Zhuang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Qi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Duan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Xi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Zhu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Xiong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>He</surname>
          </string-name>
          ,
          <string-name>
            <surname>A Comprehensive</surname>
          </string-name>
          <article-title>Survey on Transfer Learning</article-title>
          ,
          <source>Proceedings of the IEEE</source>
          <volume>109</volume>
          (
          <year>2021</year>
          )
          <fpage>43</fpage>
          -
          <lpage>76</lpage>
          . doi:
          <volume>10</volume>
          .1109/ JPROC.
          <year>2020</year>
          .
          <volume>3004555</volume>
          . arXiv:
          <year>1911</year>
          .02685.
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>C.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <article-title>Transfer learning in biomedical named entity recognition: An evaluation of BERT in the PharmaCoNER task</article-title>
          ,
          <source>in: Proceedings of The 5th Workshop on BioNLP Open</source>
          Shared Tasks, Association for Computational Linguistics, Hong Kong, China,
          <year>2019</year>
          , pp.
          <fpage>100</fpage>
          -
          <lpage>104</lpage>
          . URL: https://www.aclweb.org/anthology/D19-5715. doi:
          <volume>10</volume>
          .18653/v1/
          <fpage>D19</fpage>
          -5715.
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Peng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Yan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Lu</surname>
          </string-name>
          ,
          <article-title>Transfer Learning in Biomedical Natural Language Processing: An Evaluation of BERT and ELMo on Ten Benchmarking Datasets</article-title>
          , arXiv (
          <year>2019</year>
          )
          <fpage>58</fpage>
          -
          <lpage>65</lpage>
          . doi:
          <volume>10</volume>
          .18653/v1/w19-
          <fpage>5006</fpage>
          . arXiv:
          <year>1906</year>
          .05474.
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <surname>C.-L. Liu</surname>
            , T.-Y. Hsu,
            <given-names>Y.-S.</given-names>
          </string-name>
          <string-name>
            <surname>Chuang</surname>
          </string-name>
          , H.-y. Lee,
          <article-title>What makes multilingual BERT multilingual?</article-title>
          ,
          <source>ArXiv</source>
          (
          <year>2020</year>
          ). URL: http://arxiv.org/abs/
          <year>2010</year>
          .10938. arXiv:
          <year>2010</year>
          .10938.
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>N.</given-names>
            <surname>Nandhakumar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Sherkat</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E. E.</given-names>
            <surname>Milios</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Gu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Butler</surname>
          </string-name>
          ,
          <article-title>Clinically significant information extraction from radiology reports</article-title>
          ,
          <source>DocEng 2017 - Proceedings of the 2017 ACM Symposium on Document Engineering</source>
          (
          <year>2017</year>
          )
          <fpage>153</fpage>
          -
          <lpage>162</lpage>
          . doi:
          <volume>10</volume>
          .1145/3103010.3103023.
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>C. P.</given-names>
            <surname>Langlotz</surname>
          </string-name>
          ,
          <article-title>RadLex: A new method for indexing online educational materials</article-title>
          ,
          <source>Radiographics</source>
          <volume>26</volume>
          (
          <year>2006</year>
          )
          <fpage>1595</fpage>
          -
          <lpage>1597</lpage>
          . doi:
          <volume>10</volume>
          .1148/rg.266065168.
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <surname>J. M. Steinkamp</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Chambers</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Lalevic</surname>
            ,
            <given-names>H. M.</given-names>
          </string-name>
          <string-name>
            <surname>Zafar</surname>
            ,
            <given-names>T. S.</given-names>
          </string-name>
          <string-name>
            <surname>Cook</surname>
          </string-name>
          ,
          <article-title>Toward Complete Structured Information Extraction from Radiology Reports Using Machine Learning</article-title>
          ,
          <source>Journal of Digital Imaging</source>
          <volume>32</volume>
          (
          <year>2019</year>
          )
          <fpage>554</fpage>
          -
          <lpage>564</lpage>
          . doi:
          <volume>10</volume>
          .1007/s10278-019-00234-y.
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>S.</given-names>
            <surname>Hochreiter</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Schmidhuber</surname>
          </string-name>
          ,
          <article-title>Lstm can solve hard long time lag problems</article-title>
          ,
          <source>in: Proceedings of the 9th International Conference on Neural Information Processing Systems</source>
          , NIPS'96, MIT Press, Cambridge, MA, USA,
          <year>1996</year>
          , p.
          <fpage>473</fpage>
          -
          <lpage>479</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>A.</given-names>
            <surname>Névéol</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Dalianis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Velupillai</surname>
          </string-name>
          , G. Savova,
          <string-name>
            <surname>P. Zweigenbaum,</surname>
          </string-name>
          <article-title>Clinical Natural Language Processing in languages other than English: Opportunities and challenges</article-title>
          ,
          <source>Journal of Biomedical Semantics</source>
          <volume>9</volume>
          (
          <year>2018</year>
          )
          <fpage>1</fpage>
          -
          <lpage>13</lpage>
          . doi:
          <volume>10</volume>
          .1186/s13326-018-0179-8.
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>V.</given-names>
            <surname>Cotik</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Filippo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Roller</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Uszkoreit</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <article-title>Annotation of entities and relations in Spanish radiology reports</article-title>
          ,
          <source>International Conference Recent Advances in Natural Language Processing, RANLP 2017-Septe</source>
          (
          <year>2017</year>
          )
          <fpage>177</fpage>
          -
          <lpage>184</lpage>
          . doi:
          <volume>10</volume>
          .26615/
          <fpage>978</fpage>
          -954-452-049-6-025.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>