<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>NLNDE: The Neither-Language-Nor-Domain-Experts' Way of Spanish Medical Document De-Identi cation</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Lukas Lange</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Heike Adel</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Jannik Strotgen</string-name>
          <email>Jannik.Stroetgeng@de.bosch.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Bosch Center for Arti cial Intelligence Robert-Bosch-Campus 1</institution>
          ,
          <addr-line>71272 Renningen, Germany https://</addr-line>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Saarbrucken Graduate School of Computer Science Saarland Informatics Campus, Saarland University</institution>
          ,
          <addr-line>Saarbrucken</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Spoken Language Systems</institution>
          ,
          <addr-line>LSV</addr-line>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2019</year>
      </pub-date>
      <fpage>671</fpage>
      <lpage>678</lpage>
      <abstract>
        <p>Natural language processing has huge potential in the medical domain which recently led to a lot of research in this eld. However, a prerequisite of secure processing of medical documents, e.g., patient notes and clinical trials, is the proper de-identi cation of privacy-sensitive information. In this paper, we describe our NLNDE system, with which we participated in the MEDDOCAN competition, the medical document anonymization task of IberLEF 2019. We address the task of detecting and classifying protected health information from Spanish data as a sequence-labeling problem and investigate di erent embedding methods for our neural network. Despite dealing in a non-standard language and domain setting, the NLNDE system achieves promising results in the competition.</p>
      </abstract>
      <kwd-group>
        <kwd>De-Identi cation dings</kwd>
        <kwd>Recurrent Neural Networks</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        The anonymization of privacy-sensitive information is of increasing importance
in the age of digitalization and machine learning. It is, in particular, relevant for
texts from the medical domain that contain a large number of sensitive
information by nature. The shared task MEDDOCAN (Medical Document
Anonymization) [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] aims at automatically detecting protected health information (PHI)
from Spanish medical documents. Following the past de-identi cation task on
English PubMed abstracts [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ], it is the rst competition on this topic on Spanish
data.
      </p>
      <p>In this paper, we describe our submissions to MEDDOCAN and their
results. We, as Neither Language Nor Domain Experts (NLNDE), address the
anonymization task as a sequence-labeling problem and use a combination of</p>
    </sec>
    <sec id="sec-2">
      <title>L. Lange et al.</title>
      <p>di erent state-of-the-art approaches from natural language processing to tackle
its challenges.</p>
      <p>
        We train recurrent neural networks with conditional random eld output
layers which are state of the art for di erent sequence labeling tasks, such as named
entity recognition [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ], part-of-speech tagging [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] or de-identi cation [
        <xref ref-type="bibr" rid="ref10 ref8">8,10</xref>
        ].
Recently, the eld of natural language processing has seen another boost in
performance by using context-aware language representations which are pre-trained
on a large amount of unlabeled corpora [
        <xref ref-type="bibr" rid="ref1 ref12 ref4">1,4,12</xref>
        ]. Therefore, we experiment with
FLAIR embeddings for Spanish [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] to represent the input of our networks. In
our di erent runs, we further explore the advantages of domain-speci c fastText
embeddings [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] that have been pre-trained on SciELO and Wikipedia articles
[
        <xref ref-type="bibr" rid="ref13">13</xref>
        ].
      </p>
      <p>From a natural-language-processing perspective, the MEDDOCAN task is
interesting due to the non-standard domain (medicine) and language (Spanish) of
the documents. The results of our submissions show that state-of-the-art
architectures for sequence-labeling tasks can be directly transferred to these settings
and that domain-speci c embeddings are helpful but not necessary.
2</p>
      <sec id="sec-2-1">
        <title>Methods</title>
        <p>In this section, we rst give an overview of the di erent embedding methods we
use in our system. Second, we describe the architecture of our system.
Character embeddings
fastText</p>
        <p>FLAIR
pretrained
BiLSTM-LM
char
embeddings
BiLSTM
char
embeddings
n-gram
embeddings
&lt;añ año ños os&gt;</p>
        <p>
          Pedro de diecisiete años sin antecedentes ...
a ñ o s
… t e
a ñ o s
s i ...
Character Embedding: The characters of a word are represented by randomly
initialized embeddings. Those are passed to a bi-directional long short-term
memory network (BiLSTM). The last hidden states of the forward and backward pass
are concatenated to represent the word [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ].
        </p>
        <p>
          FastText Embedding: The fastText embeddings represent a word by the
normalized sum of the embeddings for the n-grams of the word [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ].
        </p>
        <p>
          FLAIR Embedding: FLAIR computes character-based embeddings for each word
depending on all words in the context [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ]. For this, the complete sentence is used
as the input to the BiLSTM instead of only a single word. The BiLSTM of
FLAIR is pretrained using a character-level language model objective, i.e., given
a sequence of characters, compute the probability for the next character.
2.2
        </p>
        <sec id="sec-2-1-1">
          <title>NLNDE System</title>
          <p>T1
T2</p>
          <p>NOMBRE_SUJETO_ASISTENCIA 0 5 Pedro
EDAD_SUJETO_ASISTENCIA 9 24 diecisiete años</p>
          <p>postprocessing
B-NOM_SA</p>
          <p>O</p>
          <p>B-EDAD</p>
          <p>I-EDAD</p>
          <p>O</p>
          <p>O</p>
          <p>CRF
BiLSTM
embeddings
Pedro
de
diecisiete
años
sin antecedentes ...</p>
          <p>In Figure 2, the architecture of our model is depicted. In the following, we
explain the di erent layers.</p>
          <p>
            Input Representation. We tokenize the input using the tokenizer provided by
the shared task organizers [
            <xref ref-type="bibr" rid="ref6">6</xref>
            ]. Then, we represent each token with embeddings.
In our runs, we investigate the impact of the following kinds of embeddings:
the output of an LSTM over character embeddings (50 dimensions, randomly
initialized and ne-tuned during training), domain-independent fastText
embeddings (300 dimensions, pre-trained on Spanish text [
            <xref ref-type="bibr" rid="ref5">5</xref>
            ]), domain-speci c fastText
embeddings (100 dimensions, pre-trained on Spanish SciELO and Wikipedia
articles [
            <xref ref-type="bibr" rid="ref13">13</xref>
            ]), and FLAIR embeddings (4096 dimensions, pre-trained on Spanish
text [
            <xref ref-type="bibr" rid="ref2">2</xref>
            ]). For FLAIR embeddings, we also test their pooled version (8192
dimensions, using min pooling) [
            <xref ref-type="bibr" rid="ref1">1</xref>
            ]. Note that except for the character embeddings, we
do not ne-tune any of the embeddings.
          </p>
          <p>
            BiLSTM-CRF Layers. The embeddings are fed into a BiLSTM with a
conditional random eld (CRF) output layer, similar as done by Lample et al. [
            <xref ref-type="bibr" rid="ref9">9</xref>
            ].
The CRF output layer is a linear-chain CRF, i.e., it learns transition scores
between the output classes. For training, the forward algorithm is used to sum
the scores for all possible sequences. During decoding, the Viterbi algorithm is
applied to obtain the sequence with the maximum score. Note that the
hyperparameters are the same across all runs. We use a BiLSTM hidden size of 256 and
train the network with mini-batch stochastic gradient descent using a learning
rate of 0.1 and a batch size of 32. For regularization, we employ early stopping
on the development set and apply dropout with probability 0.5 on the input
representations.
          </p>
          <p>
            Postprocessing. The output of the model is further adjusted with a
post-processing layer, similar as done by Yang et al. [
            <xref ref-type="bibr" rid="ref15">15</xref>
            ] and Liu et al. [
            <xref ref-type="bibr" rid="ref10">10</xref>
            ]. As some classes
from the annotation guidelines 4 do not occur in the training data, we tackle
them with pattern matching. For this, we use regular expressions for URLs,
IPand MAC addresses to detect the classes URL WEB and DIREC PROT INTERNET,
overwriting the results of the neural classi er.
3
          </p>
        </sec>
      </sec>
      <sec id="sec-2-2">
        <title>Submissions</title>
        <p>We submitted ve runs to the MEDDOCAN competition. All of them are based
on the architecture described in Section 2.2. They only di er in the usage of
di erent input representations.</p>
        <p>S1 (Char+fastText+Domain): Our rst run uses a combination of character
embeddings, domain-independent fastText embeddings as well as
domainspeci c fastText embeddings to represent tokens. The resulting
representation for each token has 450 dimensions.</p>
        <p>S2 (FLAIR+fastText ): In contrast to all other runs, the second run uses only
domain-independent embeddings, i.e., embeddings that have been trained
on standard narrative and news data from Common Crawl and Wikipedia.
In particular, it uses a combination of domain-independent fastText
embeddings and Flair embeddings.</p>
        <p>S3 (FLAIR+fastText+Domain ): The third run adds domain-speci c fastText
embeddings to the system of the second run in order to investigate the impact
of domain knowledge.
4 http://temu.bsc.es/meddocan/index.php/annotation-guidelines/
S4 (PooledFLAIR): The fourth run is equal to the third run, except that we use
the minimum-pooling version of the FLAIR embeddings.</p>
        <p>S5 (Ensemble): The fth run is an ensemble of the previous four runs using
weighted voting: Each classi er Ci is assigned a weight wi 2 [0:5; 3]. For each
output label, the weights of the classi ers predicting it are summed. Then,
the label with the highest score is chosen if it exceeds a speci c threshold t 2
[1; 5], or O (no PHI class) otherwise. The weights and threshold are selected
based on results on the development set as follows: w1 = 0:5, w2 = 2:0,
w3 = 2:5, w4 = 0:5 and t = 3. With these parameters, a label needs votes
from at least two classi ers (wi &lt; t; i 2 f1; 2; 3; 4g). However, the models of
the submissions S2 and S3 are assigned higher weights than S1 and S4. This
re ects their performance (see next section).
4</p>
      </sec>
      <sec id="sec-2-3">
        <title>Results and Analysis</title>
        <p>
          This section describes our results and analysis. We report the results on the
MEDDOCAN test set using the o cial shared task evaluation measures [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ].
4.1
        </p>
        <p>Results for Task 1: NER O set and Entity Type Classi cation
In the rst sub-task, the systems need to nd spans for de-identi cation and
categorize them into one of 29 classes. Table 1 presents our results on this
subtask.</p>
        <p>While the domain-independent system (run 2 with FLAIR and
domainindependent fastText embeddings) leads to the highest recall values, the third
run that also uses domain-speci c fastText embeddings achieves the highest F1
scores. This shows that integrating domain knowledge into the token
representation is bene cial. However, the di erences among the ve runs are rather small,
indicating that the architecture itself is already strong enough for the given
dataset and the impact of di erent input representations is minor.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>L. Lange et al.</title>
      <p>Since the o cial evaluation measure for this task is the strict one, we focus
our explanation on Table 2. The main ranking of our models is the same as
the ranking for sub-task 1: the addition of domain-speci c input representations
performs best. Interestingly, the domain-speci c input representations (run 3)
now perform best in terms of recall as well while the domain-independent input
representations (run 2) perform best in terms of precision.</p>
      <p>In both sub-tasks, FLAIR embeddings outperform standard character
embeddings (except for the evaluation type merge in Table 3). Also, for both
subtasks, pooling of FLAIR embeddings leads to worse results. Surprisingly, run 5,
i.e., the ensemble of the models from runs 1{4, does not improve the results over
single models.
4.3</p>
      <sec id="sec-3-1">
        <title>Confusion Matrix Analysis</title>
        <p>Table 4 shows the confusion matrix of our best performing system (run 3). It is
similar to the identity matrix, i.e., confusions between classes happen very rarely.
The most confusions happen with O, the label we assign to all non-PHI terms
which might be caused by the high number of occurrences of this class in the
training dataset. Confusions among PHI-classes happen mostly between related
classes. For example, Hospital (HOS) and Institution (INST) are confused quite
often, as Hospital is a subclass of Institution and other medical institutions
are tagged with Hospital and vice versa, e.g., Clinica Gnation is an institution
2 Abbreviations for entity types:</p>
        <p>ECDAALLDESUJET O(CAASLILSET)E,NCIA C(EEDNATDR)O, SFAALMUIDLIARES (SCUSJ)E,TO ASCISOTRERNECOIAEL(EFACTMR),ONFIECCOHAS (F(EMCAHIAL)),,
IIHDDOETSPMITIPTULALELAOCIPO(EHNROSPSOE)NR,ASOLINSDAAALNSISTEAAGNRUIITROAARMIIOENTO(ID(IED(IPDTS)PA,SS)),, IDIDICNOSSUNTJTITEAUTCCOTIOOANSAISSITSETNECN(IICANISATL), (I D(IDCNOOSNAM))-,,
BRE PERSONAL SANITARIO (NOM PS), NOMBRE SUJETO ASISTENCIA (NOM SA),
NUMERO FAX (#FAX), NUMERO TELEFONO (#TEL), OTROS SUJETO ASISTENCIA (OTRO), PAIS (PAIS),
PROFESION (PROF), SEXO SUJETO ASISTENCIA (SEXO), TERRITORIO (TER)
tagged as a hospital. Analogously, Streets (CALLE) and Territoriums (TER)
are getting confused often, as both classes are related and typically constitute
of multiple tokens. In contrast to this, Countries (PAIS) are tagged correctly
almost every time, as there is only a very limited number of countries and they
are usually single token expressions.
As mentioned above, the performance di erence between our systems is rather
small. This may be caused by the synthetic augmentation of the MEDDOCAN
data which was used to extend the texts with header and footer information
containing many PHI terms. In fact, 85% of PHI terms appear in the augmented
text parts. While this extension is necessary to cover more classes and PHI terms,
the synthetic nature of these extensions may have an impact on the performance
of automatic classi ers. Therefore, we perform a case study in which we remove
these parts from the test set and compare only the predictions found in the real
text. Only 838 out of 5661 (14.8%) annotations and only 13 out of 29 classes
remain in this experiment. The performances of our systems are decreased to F1
scores around 0.90 which is still rather high. This shows that our systems have
learned more than just to reproduce the synthetic data augmentation. However,
the performance di erences among our systems are still small, indicating that
the data augmentation was not the reason for this behavior. Note, however, that
we did not retrain our models without the synthetic augmentation.
5</p>
        <sec id="sec-3-1-1">
          <title>Conclusions</title>
          <p>In this paper, we described the system with which we participated in the
MEDDOCAN competition on automatically detecting protected health information
from Spanish medical documents. As neither language nor domain experts, we
addressed the task with a sequence labeling model. In particular, we trained
a bi-directional long short-term memory network and explored di erent input
representations. All of our runs achieved high performance with F1 scores about
97%.</p>
        </sec>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Akbik</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bergmann</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vollgraf</surname>
          </string-name>
          , R.:
          <article-title>Pooled contextualized embeddings for named entity recognition</article-title>
          .
          <source>In: Proc. of NAACL</source>
          . pp.
          <volume>724</volume>
          {
          <issue>728</issue>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Akbik</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Blythe</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vollgraf</surname>
          </string-name>
          , R.:
          <article-title>Contextual string embeddings for sequence labeling</article-title>
          .
          <source>In: Proc. of COLING</source>
          . pp.
          <volume>1638</volume>
          {
          <issue>1649</issue>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Bojanowski</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Grave</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Joulin</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mikolov</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          :
          <article-title>Enriching word vectors with subword information</article-title>
          .
          <source>Transactions of the Association for Computational Linguistics</source>
          <volume>5</volume>
          ,
          <issue>135</issue>
          {
          <fpage>146</fpage>
          (
          <year>2017</year>
          ). https://doi.org/10.1162/tacl a 00051
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Devlin</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chang</surname>
            ,
            <given-names>M.W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lee</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Toutanova</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          : BERT:
          <article-title>Pre-training of deep bidirectional transformers for language understanding</article-title>
          .
          <source>In: Proc. of NAACL</source>
          . pp.
          <volume>4171</volume>
          {
          <issue>4186</issue>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Grave</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bojanowski</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gupta</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Joulin</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mikolov</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          :
          <article-title>Learning word vectors for 157 languages</article-title>
          .
          <source>In: Proc. of LREC</source>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Intxaurrondo</surname>
            ,
            <given-names>A.:</given-names>
          </string-name>
          <article-title>SPACCC (spanish clinical case corpus) tokenizer (</article-title>
          <year>Mar 2019</year>
          ). https://doi.org/10.5281/zenodo.2586978
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Kemos</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Adel</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          , Schutze, H.:
          <article-title>Neural semi-Markov conditional random elds for robust character-based part-of-speech tagging</article-title>
          .
          <source>In: Proc. of NAACL</source>
          . pp.
          <volume>2736</volume>
          {
          <issue>2743</issue>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Khin</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Burckhardt</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Padman</surname>
            ,
            <given-names>R.:</given-names>
          </string-name>
          <article-title>A deep learning architecture for de-identi cation of patient notes: Implementation and evaluation</article-title>
          . CoRR abs/
          <year>1810</year>
          .01570 (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Lample</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ballesteros</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Subramanian</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kawakami</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dyer</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>Neural architectures for named entity recognition</article-title>
          .
          <source>In: Proc. of NAACL</source>
          (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Liu</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tang</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>Q.</given-names>
          </string-name>
          :
          <article-title>De-identi cation of clinical notes via recurrent neural network and conditional random eld</article-title>
          .
          <source>Journal of Biomedical Informatics</source>
          <volume>75</volume>
          ,
          <issue>S34</issue>
          {
          <fpage>S42</fpage>
          (
          <year>2017</year>
          ). https://doi.org/https://doi.org/10.1016/j.jbi.
          <year>2017</year>
          .
          <volume>05</volume>
          .023
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Marimon</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gonzalez-Agirre</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Intxaurrondo</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rodrguez</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lopez</surname>
            <given-names>Martin</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>J.A.</given-names>
            ,
            <surname>Villegas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            ,
            <surname>Krallinger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            :
            <surname>Automatic</surname>
          </string-name>
          de
          <article-title>-identi cation of medical texts in spanish: the meddocan track, corpus, guidelines, methods and evaluation of results</article-title>
          .
          <source>In: Proceedings of the Iberian Languages Evaluation Forum (IberLEF</source>
          <year>2019</year>
          ). vol.
          <source>TBA</source>
          , p.
          <source>TBA. CEUR Workshop Proceedings (CEUR-WS.org)</source>
          , Bilbao,
          <source>Spain (Sep</source>
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Peters</surname>
            ,
            <given-names>M.E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Neumann</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Iyyer</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gardner</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Clark</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lee</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zettlemoyer</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          :
          <article-title>Deep contextualized word representations</article-title>
          .
          <source>In: Proc. of NAACL</source>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Soares</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Villegas</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gonzalez-Agirre</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Krallinger</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Armengol-Estape</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          :
          <article-title>Medical word embeddings for Spanish: Development and evaluation (</article-title>
          <year>Jun 2019</year>
          ), https://www.aclweb.org/anthology/W19-1916
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Stubbs</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Uzuner</surname>
            ,
            <given-names>O</given-names>
          </string-name>
          .
          <article-title>: Annotating longitudinal clinical narratives for deidenti cation: The 2014 i2b2/uthealth corpus</article-title>
          .
          <source>Journal of biomedical informatics 58, S20{S29</source>
          (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Yang</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Garibaldi</surname>
            ,
            <given-names>J.M.:</given-names>
          </string-name>
          <article-title>Automatic detection of protected health information from clinic narratives</article-title>
          .
          <source>Journal of Biomedical Informatics</source>
          <volume>58</volume>
          ,
          <issue>S30</issue>
          {
          <fpage>S38</fpage>
          (
          <year>2015</year>
          ). https://doi.org/https://doi.org/10.1016/j.jbi.
          <year>2015</year>
          .
          <volume>06</volume>
          .015
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>