=Paper=
{{Paper
|id=Vol-2421/MEDDOCAN_paper_3
|storemode=property
|title=VSP at MEDDOCAN 2019 De-Identification of Medical Documents in Spanish with Recurrent Neural Networks
|pdfUrl=https://ceur-ws.org/Vol-2421/MEDDOCAN_paper_3.pdf
|volume=Vol-2421
|authors=Víctor Suárez-Paniagua
|dblpUrl=https://dblp.org/rec/conf/sepln/Suarez-Paniagua19a
}}
==VSP at MEDDOCAN 2019 De-Identification of Medical Documents in Spanish with Recurrent Neural Networks==
VSP at MEDDOCAN 2019 De-Identification of Medical Documents in Spanish with Recurrent Neural Networks Vı́ctor Suárez-Paniagua Computer Science Department, Carlos III University of Madrid. Leganés 28911, Madrid, Spain. vspaniag@inf.uc3m.es http://hulat.inf.uc3m.es/en/nosotros/miembros/vsuarez Abstract. This work presents the participation in the MEDDOCAN Task of the VSP team with a neural model for the Named Entity Recog- nition of medical documents in Spanish. The Neural Network consists of a two-layer model that creates a feature vector for each word of the sen- tences. The first layer uses the character information of each word and the output is aggregated to the second layer together with its word em- bedding in order to create the feature vector of the word. Both layers are implemented with a bidirectional Recurrent Neural Network with LSTM cells. Moreover, a Conditional Random Field layer classifies the word vectors in one of the 29 types of Protected Health Information (PHI). The system obtains a performance of 86.01%, 87.03%, and 89,12% in F1 for the classification of the entity types, the sensitive span detection, and both tasks merged, respectively. The model shows very high and promising results being a basic approach without using pretrained word embeddings or any hand-crafted feature. Keywords: Named Entity Recognition, Deep Learning, Recurrent Neu- ral Network, Medical Documents 1 Introduction Nowadays, healthcare professionals deal with a high amount of unstructured documents that makes very difficult the task of finding the essential data in medical documents. Decreasing the time-consuming task of retrieving the most relevant information can help the fastness of generating a diagnosis for patients by doctors. Instead the vast of information are available as Electronic Health Record (EHR), the manual annotation of them is impracticable because the highly increasing number of generated documents per day and also because they contain sensitive data and Protected Health Information (PHI). For this reason, the development of an automatic system that identifies sensitive information from medical documents is vital for helping doctors and preserving patient con- fidentiality. The i2b2 shared task was the first Natural Language Processing (NLP) chal- lenge for identifying PHI in the clinical narratives [13]. The second edition of the Copyright c 2019 for this paper by its authors. Use permitted under Creative Com- mons License Attribution 4.0 International (CC BY 4.0). IberLEF 2019, 24 Septem- ber 2019, Bilbao, Spain. Proceedings of the Iberian Languages Evaluation Forum (IberLEF 2019) Vı́ctor Suárez-Paniagua i2b2 shared task Track 1 [12] created a gold standard dataset with annotations of the PHI categories from 1,304 medical records in English. In this competition, the highest ranking system used the Conditional Random Field (CRF) classifier together with hand-written rules for the de-identification of clinical narratives obtaining very promising results with 97.68% in F1 [14]. The goal of the Iberian Languages Evaluation Forum (IberLEF) 2019, which includes the TASS and IberEval workshops, is to create NLP challenges using corpora written in one of the Iberian languages (Spanish, Portuguese, Catalan, Basque or Galician). Following the i2b2 de-identification task, the Medical Doc- ument Anonymization task (MEDDOCAN) encourages the research community to design NLP systems for the identification of PHI from clinical texts in Spanish [9]. For this purpose, a corpus of 1,000 clinical case studies with PHI phrases was manually annotated by health documentalists. Currently, Deep Learning approaches overcome traditional machine learning systems on the majority of NLP tasks, such as text classification [6], language modeling [10] and machine translation [1]. Moreover, these models have the ad- vantage of automatically learn the most relevant features without defining rules by hand. Concretely, the state-of-the-art performance for Named Entity Recog- nition (NER) task is an LSTM-CRF Model proposed by [8]. The main idea of this system is to create a word vector representation using a bidirectional Re- current Neural Network with LSTM cells (BiLSTM) with character information encoded in another BiLSTM layer in order to classify the tag of each word in the sentences with a CRF classifier. Following this approach, the system proposed in [3] uses a BiLSTM-CRF Model with character and word levels for the de- identification of patient notes using the i2b2 dataset. This approach overcomes the top ranking system in this task reaching to 97.88% in F1. This paper presents the participation of the VSP team at the tasks proposed by MEDDOCAN about the classification of PHI types and the sensitive span detection from medical documents in Spanish. The proposed system follows the same approaches of [8] and [3] with some modifications for the Spanish language implemented with NeuroNER tool [2]. 2 Dataset The corpus of the MEDDOCAN task contains 1,000 clinical cases with PHI enti- ties manually annotated by health documentalists. The documents are randomly divided into the training, validation and test sets for creating, developing and ranking the different systems, respectively. Similarly to the annotation schema of the i2b2 de-identification tasks, the named entities are annotated according to their offsets and their type for each de- tection and classification (see Figure 1). The 29 types of the annotated PHI men- tions follow the Health Insurance Portability and Accountability Act (HIPAA) guidelines for Spanish the health records aggregating some PHI entities. 655 Proceedings of the Iberian Languages Evaluation Forum (IberLEF 2019) VSP at MEDDOCAN 2019 Fig. 1. The annotated offsets and types of the PHI entities in the sentence ’Informe clı́nico del paciente: Adolescente Varón de diecisiete años.’. English translation: ’Clin- ical report of the patient: male teenager of seventeen years.’ 3 Neural model This section presents the Neural architecture for the classification of the PHI entity types and the sensitive span detection using medical documents in Span- ish. Figure 2 shows the entire process of the model using two BiLSTMs for the character and token levels in order to create each word representation until its classification by a CRF. Fig. 2. Neural model for the de-identification of Medical Documents in Spanish using the MEDDOCAN task 2019 corpus. 656 Proceedings of the Iberian Languages Evaluation Forum (IberLEF 2019) Vı́ctor Suárez-Paniagua 3.1 Data preprocessing Before using the system, the documents of the corpus are preprocessed in order to prepare the inputs for the Neural model. Firstly, the clinical cases are separated into sentences using a sentence splitter and the words of these sentences are extracted by a tokenizer, both were adapted for the Spanish language. Once the sentences are divided into word, the BIOES tag schema encodes each token with an entity type. The tag B defines the beginning token of a mention, the I tag defines the inside token of a mention, the E tag defines the ending token of a mention, the S tag indicates that the mention has a single token and the O tag indicates the outside tokens that do not belong to any mention. In many previous NER tasks, using this codification is better than the BIO tag scheme [11], but the number of labels increases because there are two additional tags for each class. Thus, the number of possible classes are the 4 tags times the 29 PHI classes and the O tag for the MEDDOCAN corpus. For the experiments, all the previous processes are performed by the spaCy tool in Python [4]. 3.2 BiLSTM layers RNNs are very effective in feature learning when the inputs are sequences. This Deep Learning model uses two different weights for the input and for the previous output as: h(t) = f (Wx(t) + Uh(t − 1) + b) where h(t) is the output at t time of the input x, f is a non-linear function, W are the weights for the current input, U are the weights for the previous output, and b the bias term of the Neural Network. However, the basic RNN cannot capture the long dependencies because it loses the information of the gradients as long as the back-propagation is applied to the previous states. For this reason, the incorporation of cell units into the RNN computation solves the long propagation of the gradient problem. The Long Short-Term Memory cell (LSTM) [5] defines four gates for creating a word representation taking the information of the current and previous cells. The input gate it , the forget gate ft and the output gate ot for the current t step transform the input vector xt taking the previous output ht−1 using its corresponding weights and bias computed with a sigmoid function. The cell state ct takes the information given from the previous cell state ct−1 regulated by the forget cell and the information given from the current cell c0t regulated by the input cell using the element-wise represented as: 657 Proceedings of the Iberian Languages Evaluation Forum (IberLEF 2019) VSP at MEDDOCAN 2019 ft = σ(Wf · [ht−1 , xt ] + bf ) it = σ(Wi · [ht−1 , xt ] + bi ) 0 ct = tanh(Wc · [ht−1 , xt ] + bc ) ct = ft ∗ ct−1 + it ∗ c0t ot = σ(Wo · [ht−1 , xt ] + bo ) ht = ot ∗ tanh(ct ) Finally, the current output ht is represented with the hyperbolic function of the cell state and controlled by the output gate. Furthermore, another LSTM can be applied in the other direction from the end of the sequence to the start. Com- puting the two representations is beneficial for extracting the relevant features of each word because they have dependencies in both directions. Character level The first layer takes each word of the sentences individually. These tokens are decomposed into characters that are the input of the BiLSTM. Once all the inputs are computed by the network, the last output vectors of both directions are concatenated in order to create the vector representation of the word according to its characters. Token level The second layer takes the embedding of each word in the sentence and concatenates them with the outputs of the first BiLSTM with the character representation. In addition, a Dropout layer is applied to the word representation in order to prevent overfitting in the training phase. In this case, the outputs of each direction in one token are concatenated for the classification layer. 3.3 Contional Random Field Classifier CRF [7] is the sequential version of the Softmax that aggregates the label pre- dicted in the previous output as part of the input. In NER tasks, CRF shows better results than Softmax because it adds a higher probability to the correct labelled sequence. For instance, the I tag cannot be before a B tag or after a E tag by definition. For the proposed system, the CRF classifies the output vector of the BiLSTM layer with the token information in one of the classes. 4 Results and Discussion The architecture was trained over the training set during 25 epochs with shuffled mini-batches and choosing the best performance over the validation set. The values of the two BiLSTM and CRF parameters for generating the prediction of the test set are presented in Table 1. The embeddings of the characters and words are randomly initialized and learned during the training of the network. 658 Proceedings of the Iberian Languages Evaluation Forum (IberLEF 2019) Vı́ctor Suárez-Paniagua Table 1. The parameters of the Neural model and their values used for the MEDDO- CAN results. Parameter Value Character embeddings dimension 25 Character-level LSTM hidden units 25 Word embeddings dimension 300 Word-level LSTM hidden units 256 Optimizer Adam Learning rate 0.001 Dropout rate 0.5 Gradient clipping 5 Additionally, a gradient clipping keeps the weight of the network in a low range preventing the exploding gradient problem. The results were measured with precision (P), recall (R) and F-measure (F1) using the True Positives (TP), False Positives (FP) and False Negatives (FN) for its calculation. Table 2 presents the results of the Neural Model with the two BiLSTM levels and the CRF classifier over the test set of the MEDDOCAN task. The performance over the NER offset and entity type classification (Task 1) shows an 86,01% in F1 and the performance over the sensitive token detection (Task 2) shows an 87,03% in F1 taking into consideration only if the entities have exact boundary match and entity type (Strict). Thus, the results for both tasks merged reach to 89,12% in F1. From the table, it can be observed that the number of FN and FP are very similar giving very similar Precision and Recall results in all the classes. On the one hand, there are classes with very high performance, such as CORREO ELECTRONICO, EDAD SUJETO ASISTENCIA, FECHAS, NOMBRE SUJETO ASISTENCIA and PAIS that are greater than the 95% in F1 because of the data is presented in the same location between documents and they are easy to disambiguate from the remaining classes. On the other hand, the classes of OTROS SUJETO ASISTENCIA and PROFESION shows a very low performance because they have a very small number of instances in the training set making hard the learning of their rep- resentation in the network. In order to alleviate this problem, the use of over- sampling techniques is proposed to increase the number of instances of the less representative classes and making more balanced this dataset. 5 Conclusions and Future work This work proposes a Neural model for the detection and classification of PHI from clinical texts in Spanish. The architecture is based on RNNs in both direc- tion of the sentences using LSTM for the computation of the outputs. Finally, a CRF classifier performs the classification for tagging the PHI entity types. The results shows a performance of 86.01% and 87.03% in F1 for the classifi- 659 Proceedings of the Iberian Languages Evaluation Forum (IberLEF 2019) VSP at MEDDOCAN 2019 Table 2. Results of the Neural Model over the test set of the MEDDOCAN. Label TP FN FP R P F1 CALLE 226 187 225 54,72% 50,11% 52,31% CENTRO SALUD 4 2 3 66,67% 57,14% 61,54% CORREO ELECTRONICO 244 5 7 97,99% 97,21% 97,6% EDAD SUJETO ASISTENCIA 504 14 37 97,3% 93,16% 95,18% FAMILIARES SUJETO ASISTENCIA 55 26 44 67,9% 55,56% 61,11% FECHAS 585 26 25 95,74% 95,9% 95,82% HOSPITAL 102 28 35 78,46% 74,45% 76,4% ID ASEGURAMIENTO 184 14 22 92,93% 89,32% 91,09% ID CONTACTO ASISTENCIAL 35 4 4 89,74% 89,74% 89,74% ID SUJETO ASISTENCIA 251 32 37 88,69% 87,15% 87,92% ID TITULACION PERSONAL SANITARIO 217 17 19 92,74% 91,95% 92,34% INSTITUCION 29 38 31 43,28% 48,33% 45,67% NOMBRE PERSONAL SANITARIO 468 33 26 93,41% 94,74% 94,07% NOMBRE SUJETO ASISTENCIA 502 0 4 100% 99,21% 99,6% NUMERO FAX 5 2 2 71,43% 71,43% 71,43% NUMERO TELEFONO 20 6 3 76,92% 86,96% 81,63% OTROS SUJETO ASISTENCIA 0 7 1 0% 0% 0% PAIS 350 13 5 96,42% 98,59% 97,49% PROFESION 2 7 5 22,22% 28,57% 25% SEXO SUJETO ASISTENCIA 228 233 232 49,46% 49,57% 49,51% TERRITORIO 885 71 61 92,57% 93,55% 93,06% Task 1 4896 765 828 86,49% 85,53% 86,01% Task 2 (Strict) - - - 87,51% 86,55% 87,03% Task 2 (Merged) - - - 89,36% 88,88% 89,12% cation of the entity types and the sensitive span detection over the MEDDO- CAN corpus giving 89,12% in F1 for the merged tasks as the official result. The results are very similar in Precision and Recall for all the classes giving a low performance in the less representative classes and a higher performance in the well-structured PHI entities, such as NOMBRE SUJETO ASISTENCIA EDAD SUJETO ASISTENCIA, CORREO ELECTRONICO, FECHAS, and PAIS. As future work, exploring the contribution of each representation individually and fine-tuning the parameters of the model will be useful in order to increase the performance. In addition, the aggregation of embeddings from different external information, such as Part-of-Speech tags, syntactic parse trees or semantic tags, could increase the representation of each word for improving its classification. Moreover, the sentence splitter of spaCy seems to divide sentences when some acronyms appear, such as ’Dr.’, ’Dra.’, ’Sr.’ or ’Sra.’ (Spanish honorific prefix). For this reason, the creation of simple rules in order to avoid these cases could be beneficial for increasing the performance. Furthermore, adding more layers to each BiLSTM is proposed to be included in the architecture. 660 Proceedings of the Iberian Languages Evaluation Forum (IberLEF 2019) Vı́ctor Suárez-Paniagua References 1. Cho, K., van Merrienboer, B., Gulcehre, C., Bahdanau, D., Bougares, F., Schwenk, H., Bengio, Y.: Learning phrase representations using RNN encoder–decoder for statistical machine translation. In: Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP). pp. 1724–1734. Association for Computational Linguistics, Doha, Qatar (Oct 2014). https://doi.org/10.3115/v1/D14-1179, https://www.aclweb. org/anthology/D14-1179 2. Dernoncourt, F., Lee, J.Y., Szolovits, P.: NeuroNER: an easy-to-use program for named-entity recognition based on neural networks. In: Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing: System Demonstrations. pp. 97–102. Association for Computational Linguistics, Copen- hagen, Denmark (Sep 2017). https://doi.org/10.18653/v1/D17-2017, https:// www.aclweb.org/anthology/D17-2017 3. Dernoncourt, F., Young Lee, J., Uzuner, O., Szolovits, P.: De-identification of pa- tient notes with recurrent neural networks. Journal of the American Medical Infor- matics Association : JAMIA 24 (06 2016). https://doi.org/10.1093/jamia/ocw156 4. Explosion AI: spaCy - Industrial-strength Natural Language Processing in Python (2017), https://spacy.io/ 5. Hochreiter, S., Schmidhuber, J.: Long short-term memory. Neural computation 9(8), 1735–1780 (1997) 6. Kim, Y.: Convolutional neural networks for sentence classification. In: Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP). pp. 1746–1751 (2014) 7. Lafferty, J.D., McCallum, A., Pereira, F.C.N.: Conditional random fields: Prob- abilistic models for segmenting and labeling sequence data pp. 282–289 (2001), http://dl.acm.org/citation.cfm?id=645530.655813 8. Lample, G., Ballesteros, M., Subramanian, S., Kawakami, K., Dyer, C.: Neural architectures for named entity recognition. In: Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. pp. 260–270. Association for Computational Lin- guistics, San Diego, California (Jun 2016). https://doi.org/10.18653/v1/N16-1030, https://www.aclweb.org/anthology/N16-1030 9. Marimon, M., Gonzalez-Agirre, A., Intxaurrondo, A., Rodrı́guez, H., Lopez Mar- tin, J.A., Villegas, M., Krallinger, M.: Automatic de-identification of medical texts in spanish: the meddocan track, corpus, guidelines, methods and evaluation of re- sults. In: Proceedings of the Iberian Languages Evaluation Forum (IberLEF 2019). vol. TBA, p. TBA. CEUR Workshop Proceedings (CEUR-WS.org), Bilbao, Spain (Sep 2019), TBA 10. Mikolov, T., Sutskever, I., Chen, K., Corrado, G.S., Dean, J.: Distributed repre- sentations of words and phrases and their compositionality. In: Advances in neural information processing systems. pp. 3111–3119 (2013) 11. Ratinov, L., Roth, D.: Design challenges and misconceptions in named entity recog- nition. In: Proceedings of the Thirteenth Conference on Computational Natural Language Learning (CoNLL-2009). pp. 147–155. Association for Computational Linguistics (2009), http://aclweb.org/anthology/W09-1119 12. Stubbs, A., Kotfila, C., Uzuner, O.: Automated systems for the de- identification of longitudinal clinical narratives: Overview of 2014 i2b2/uthealth shared task track 1. Journal of biomedical informatics 58 (07 2015). https://doi.org/10.1016/j.jbi.2015.06.007 661 Proceedings of the Iberian Languages Evaluation Forum (IberLEF 2019) VSP at MEDDOCAN 2019 13. Özlem Uzuner, Luo, Y., Szolovits, P.: Evaluating the state-of-the-art in automatic de-identification. Journal of the American Medical Informatics Association 14(5), 550 – 563 (2007). https://doi.org/https://doi.org/10.1197/jamia.M2444, http:// www.sciencedirect.com/science/article/pii/S106750270700179X 14. Yang, H., Garibaldi, J.M.: Automatic detection of protected health information from clinic narratives. Journal of Biomedical Informatics 58, S30–S38 (2015) 662