<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Transformer-Based Disease and Drug Named Entity Recognition in Multilingual Clinical Texts: MultiCardioNER challenge Notebook for the BioASQ Lab at CLEF 2024</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Anna</forename><surname>Aksenova</surname></persName>
							<email>anna.aksenova@ontotext.com</email>
							<affiliation key="aff0">
								<orgName type="institution">Aalto University</orgName>
								<address>
									<country key="FI">Finland</country>
								</address>
							</affiliation>
							<affiliation key="aff1">
								<orgName type="institution">Ontotext</orgName>
								<address>
									<settlement>Sofia</settlement>
									<country key="BG">Bulgaria</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Aleksis</forename><surname>Datseris</surname></persName>
							<email>aleksis.datseris@ontotext.com</email>
							<affiliation key="aff1">
								<orgName type="institution">Ontotext</orgName>
								<address>
									<settlement>Sofia</settlement>
									<country key="BG">Bulgaria</country>
								</address>
							</affiliation>
							<affiliation key="aff2">
								<orgName type="department">Faculty of Mathematics and Informatics</orgName>
								<orgName type="institution">Sofia University &quot;St. Kliment Ohridski&quot;</orgName>
								<address>
									<settlement>Sofia</settlement>
									<country key="BG">Bulgaria</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Sylvia</forename><surname>Vassileva</surname></persName>
							<email>svasileva@fmi.uni-sofia.bg</email>
							<affiliation key="aff2">
								<orgName type="department">Faculty of Mathematics and Informatics</orgName>
								<orgName type="institution">Sofia University &quot;St. Kliment Ohridski&quot;</orgName>
								<address>
									<settlement>Sofia</settlement>
									<country key="BG">Bulgaria</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Svetla</forename><surname>Boytcheva</surname></persName>
							<email>svetla@uni-sofia.bg</email>
							<affiliation key="aff1">
								<orgName type="institution">Ontotext</orgName>
								<address>
									<settlement>Sofia</settlement>
									<country key="BG">Bulgaria</country>
								</address>
							</affiliation>
							<affiliation key="aff2">
								<orgName type="department">Faculty of Mathematics and Informatics</orgName>
								<orgName type="institution">Sofia University &quot;St. Kliment Ohridski&quot;</orgName>
								<address>
									<settlement>Sofia</settlement>
									<country key="BG">Bulgaria</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Transformer-Based Disease and Drug Named Entity Recognition in Multilingual Clinical Texts: MultiCardioNER challenge Notebook for the BioASQ Lab at CLEF 2024</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">4DCD5F904B24762620AA6202875AEC9C</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T17:56+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Named entity recognition (NER), Biomedical NLP, Medication extraction, Diagnosis extraction, Clinical NER, Multilingual NER (S. Boytcheva) 0000-0002-3489-874X (A. Aksenova)</term>
					<term>0000-0002-2257-0659 (S. Vassileva)</term>
					<term>0000-0002-5542-9168 (S. Boytcheva)</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>This paper presents a transformer-based approach for disease Named Entity Recognition (NER) in Spanish clinical texts using the DisTEMIST dataset and drug multi-lingual NER in Spanish, English and Italian clinical texts using the DrugTEMIST dataset. For the disease NER task, we use CLIN-X-ES, a BERT-based model pretrained on Spanish clinical texts and additional pretrained on a custom dataset, and fine-tuned on token classification, achieving F1 score 0.8049 on the test set. For the drug NER task, we experiment with language-specific clinical models as well as general domain multilingual models and achieved the best results with the language-specific models. For Spanish we fine-tuned the CLIN-X-ES model and our best model showed 0.9238 F1 score, for English we fine-tuned BioLinkBERT which scored F1 -0.9223, and for Italian we pretrained the CLIN-X-ES model with a custom Italian dataset and achieved F1 score -0.8838. Our system placed first on the English and Italian tracks in the drug subtask in MultiCardioNER challenge.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Clinical narratives are a valuable source of healthcare information, but it requires design of special NLP models for effective data extraction. Automated identification of key terms such as diseases, medications, and procedures within clinical documents is known as clinical named entity recognition (NER). This process plays a significant role in clinical natural language processing (NLP) by facilitating the extraction of structured data from clinical narratives for subsequent analysis and interpretation by downstream healthcare applications. MultiCardioNER 1 <ref type="bibr" target="#b0">[1]</ref> is a shared task part of CLEF BioASQ <ref type="bibr" target="#b1">[2]</ref>, which aims to detect diseases in Spanish clinical texts as well as drugs in a multilingual setting, including Spanish, English and Italian. The organizers have provided annotated datasets for training and evaluation of the systems -DisTEMIST for disease NER and DrugTEMIST for drug NER. The shared task consists of two subtask -Subtask 1 addressing disease recognition in Spanish, and Subtask 2 addressing drug recognition in multiple languages. In both subtasks, the challenge is to recognize terms in cardiology reports.</p><p>This paper describes our approach for disease and drug NER which we submitted for the Mul-tiCardioNER challenge. Our code is available on GitHub 2 . The contributions of this paper are as follows:</p><p>• Developed a system for disease entity recognition in Spanish and performed different experiments with BERT-based models, achieving 0.8049 F1 score on the DisTEMIST dataset; • Developed a system for drug entity recognition in Spanish, English, and Italian which achieved state-of-the-art (SOTA) results on English and Italian DrugTEMIST dataset and very competitive results on Spanish; • We investigated and compared the performance of multilingual BERT-based models vs languagespecific models for named entity recognition; • We adapted a clinical Spanish RoBERTa model to the Italian language and showed the best result on the drug NER task for Italian;</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Related Work</head><p>The state-of-the-art methods for biomedical named entity recognition are predominantly using deep learning based models <ref type="bibr" target="#b2">[3]</ref>, <ref type="bibr" target="#b3">[4]</ref>. The most recent NER approaches for clinical documents also include some hybrid models like dictionary guided attention based model <ref type="bibr" target="#b4">[5]</ref> and transfer learning <ref type="bibr" target="#b5">[6]</ref>, <ref type="bibr" target="#b6">[7]</ref>.</p><p>Besides classical approaches like machine learning <ref type="bibr" target="#b7">[8]</ref>, hidden Markov models (HMM) and conditional random fields (CRF) <ref type="bibr" target="#b8">[9]</ref>, an interesting application of Fourier Networks for NER and relation extraction were proposed in <ref type="bibr" target="#b9">[10]</ref>. Another direction of research is using models based on knowledge graphs <ref type="bibr" target="#b10">[11]</ref> as additional source for information. The limited availability of annotated data, imbalanced training datasets and lack of resources in low resource languages triggered another direction in the NER model development using data augmentation techniques to tackle these issues <ref type="bibr" target="#b11">[12]</ref>, <ref type="bibr" target="#b12">[13]</ref>. Specifically for the task of NER for medication extraction the SOTA methods are based on BiL-STM+CRF <ref type="bibr" target="#b13">[14]</ref> reporting F1-score 0.93 for the best performing system for NER tasks over MIMIC-III<ref type="foot" target="#foot_0">3</ref> dataset. Another approach is based on Question-Answering (QA) for medication event extraction <ref type="bibr" target="#b14">[15]</ref> translating the NER task to span identification task in QA, reporting NER F1-score 0.98. The classical models like SVM, CRF and rule-based models <ref type="bibr" target="#b15">[16]</ref>, <ref type="bibr" target="#b16">[17]</ref> show comparable results. In the n2c2 shared task on medication event extraction in clinical notes <ref type="bibr" target="#b17">[18]</ref> the top score model for NER task scores strict F1 0.97 using transformer based pretrained LLM, using a BERT-based model (RoBERTa-large-PM-M3-Voc) with classification layer and BILOU tags.</p><p>For the task of NER for diagnoses, the SOTA methods are also based on transformers <ref type="bibr" target="#b18">[19]</ref>, <ref type="bibr" target="#b19">[20]</ref> of various architectures. Most widely used are domain and task adaptations of the transformer architecture, such as BERT, RoBERTa and ELECTRA. Additional enhancement using knowledge bases in deep learning models <ref type="bibr" target="#b20">[21]</ref> can help in dataset annotation and expansion.</p><p>The organizers of MultiCardioNER have organized multiple challenges in the area of information extraction from Spanish clinical texts for different entity types -diseases, procedures, symptoms, etc. Transformer-based approaches are very commonly used for NER tasks in different languages. In previous challenges for diseases using the DisTEMIST dataset, the top teams have used BERT-based models trained for token classification like Spanish RoBERTa (PlanTL-GOB-ES/roberta-base-biomedicalclinical-es <ref type="bibr" target="#b21">[22]</ref>) ( <ref type="bibr" target="#b22">[23]</ref>, <ref type="bibr" target="#b23">[24]</ref>, <ref type="bibr" target="#b24">[25]</ref>), mBERT and the Spanish BETO <ref type="bibr" target="#b25">[26]</ref>. In a similar challenge for named entity recognition of medical procedures in Spanish text, competitors used an architecture including the same Spanish RoBERTa or XLM-RoBERTa model and adding BiLSTM and CRF layers on top ( <ref type="bibr" target="#b26">[27]</ref>, <ref type="bibr" target="#b27">[28]</ref>). In the task of symptoms NER, ensemble of transformer models for Spanish clinical text achieved the best result -F1 0.74 (strict) <ref type="bibr" target="#b28">[29]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Data</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.">Subtask 1</head><p>MultiCardioNER corpus for disease named entity recognition task consists of two data subsets of different nature: DisTEMIST and CardioCCC. The DisTEMIST dataset consists of 1000 clinical cases of different medical specialities (incl. oncology, otorhinolaryngology, dentistry, pediatrics, primary care, allergology, radiology, psychiatry, ophthalmology and more). CardioCCC is a collection of 508 cardiology clinical case reports, which are longer on average than the DisTEMIST reports. The CardioCCC dataset contains 508 documents, split in 258 for development and 250 for testing. Following the suggestions by the organizers, we used DisTEMIST as a train set, leaving CardioCCC documents for validation. Additionally, a custom dataset was used for clinical domain adaptation (see Section 3.2). Each team had to provide their model predictions on a dataset containing the CardioCCC test data and a large unlabelled background dataset. At the time of submission, the organizers had not released which examples were part of the test set. Therefore the number of documents in the prediction set is a lot higher than the train or validation sets, however only a small portion was used for evaluating the model performance by the organizers. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.">Subtask 2</head><p>Similar to Subtask1, the MultiCardioNER corpus for the drug prediction task consists of two data subsets of different nature: DrugTEMIST and CardioCCC. The difference between the datasets is quite substantial. DrugTEMIST is a task-specific adaptation of DisTEMIST dataset that consists of 1000 clinical cases of different medical specialities. CardioCCC is a collection of 508 cardiology clinical case reports, meaning that reports are longer on average. Similarly to subtask 1, we used DrugTEMIST as a train set, leaving CardioCCC documents for validation. One of the important aspects of the dataset is the sparsity of annotations. For DrugTEMIST only 9% sentences contain drugs, while for CardioCCC the share is even lower -6%.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3.">Custom Medical Text Dataset</head><p>To adapt foundation models for the clinical domain, we collected language-specific datasets with raw texts. The data statistics can be found in Table <ref type="table" target="#tab_1">2</ref>. The data was collected using the following sources:</p><p>1. Wikidata concepts related to medicine: We ran a SPARQL query over WikiData<ref type="foot" target="#foot_1">4</ref> extracting labels that are included in the following medical ontologies and classifications: ICD-11 <ref type="foot" target="#foot_2">5</ref> ; ICD-10, ICD-10 CM<ref type="foot" target="#foot_3">6</ref> , Symptom Ontology<ref type="foot" target="#foot_4">7</ref> , eMedicine <ref type="foot" target="#foot_5">8</ref> , DiseasesDB, MedlinePlus<ref type="foot" target="#foot_6">9</ref> , MONDO 10 , Human Disease Ontology 11 , SNOMED CT 12 , UMLS 13 . 2. Wikipedia articles related to medical concepts: Based on the extracted WikiData concepts, we went through the concept list and downloaded the texts of WikiPedia articles that were created for those concepts. For some concepts the articles did not exist. We used Mediawiki API 14 for text extraction. 3. Abbreviation lists found online: For each of the languages, we browsed for medical abbreviation lists available online. Some of them were extracted from the open-source articles, others were scraped from websites. 4. Drug descriptions: As drug descriptions usually contain quite useful information on symptoms, side effects and dosages, we leveraged multilingual drug description lists 15 . 5. EMA medical documentation: We leveraged a parallel corpus of the European Medical Agency Documentation 16 . 6. Machine-translated data: To enrich the amount of medical data, we decided to use machine translated datasets. For this purpose, we used medical abstracts corpus 17 .</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.4.">Drug Gazetteer</head><p>As drug lists are dynamic and changing over time, specifically designed dictionaries were created based on various official drug sources to support model-based drug extraction. For the English drug dictionary, we used OMOP Standardized Vocabulary V5.0 18 (incl. ATC, RxNorm), DrugCentral 19 (FDA Approved Drugs, EMA Approved Drugs, PMDA Approved Drugs, PMDA+EMA+FDA Approved Drug), DrugBank 20 , DailyMed 21 (NIHS human drugs), Top250 22 , UnatedHealthcare 23 , and Drugs.com 24 (My Med List). For Spanish, we used Centro de información online de medicamentos de la AEMPS -(CIMA) 25  (incl. ATC Spanish version and Arbol Medicamentos DSCA Spanish). For Italian, we used ATC and Lists of Class A and Class H medicinal products of Italian Medicine Agency 26 . The drug names and synonyms are aggregated in three dictionaries for English, Italian and Spanish. The total number of generic and brand names of drugs included in the cleaned lists, after removing duplication are presented on Table <ref type="table" target="#tab_2">3</ref>. In addition to the drug dictionaries were used some procedures names for lab test from LOINC 27 for all languages in order to disambiguate some drug mentions from lab tests for measuring levels of some minerals and vitamins, like Vitamin D, Magnesium, Calcium, etc. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.5.">Data pre-processing</head><p>As clinical documents are quite lengthy, especially in the case of the CardioCCC corpus, we decided to split the documents into sentences using the following tools for different languages:</p><p>• English -MedSpaCy Sentence Splitting <ref type="foot" target="#foot_7">28</ref>• Italian -Tint Sentence Splitting<ref type="foot" target="#foot_8">29</ref>  <ref type="bibr" target="#b29">[30]</ref> • Spanish -SPACCC Sentence splitter <ref type="foot" target="#foot_9">30</ref>Afterwards, we used the Brat tool<ref type="foot" target="#foot_10">31</ref>  <ref type="bibr" target="#b30">[31]</ref> for data transformation from BRAT to CONLL format. The dataset statistics after pre-processing are shown in Table <ref type="table" target="#tab_3">4</ref>. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Methods</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1.">Subtask 1</head><p>Our system was built using transformer-based models that were either multilingual or adapted to Spanish and that were preferably adapted to the biomedical domain as well. The models we used were:</p><p>• PlanTL-GOB-ES/roberta-base-biomedical-clinical-es <ref type="bibr" target="#b21">[22]</ref>: Biomedical pretrained language model for Spanish. This model is a RoBERTa-based model trained on a biomedicalclinical corpus in Spanish collected from several sources. • CLIN-X-ES <ref type="bibr" target="#b31">[32]</ref>: This model is based on the multilingual XLM-R transformer (xlm-roberta-large) further pretrained on a Spanish clinical corpus. • DeBERTa v3 <ref type="bibr" target="#b32">[33]</ref>: A transformer-based model with disentagled attention trained using ELECTRA style pretraining. Both base and large versions were used and also a version of DeBERTa that was further pretrained on clinical data. • mDeBERTa v3 <ref type="foot" target="#foot_11">32</ref> : A multilingual version of DeBERTa.</p><p>Additionally, some of the models were further pretrained on medical data from the Custom Medical Text Dataset described in Section 3.3. Approximately 0.08 of the tokens are annotated entities per sentence which corresponds to a very sparse annotation setting. Hence the majority of tokens to be evaluated by the model will be negative examples. Therefore, we tried to fine-tune our models with different class weight ratios (positive to negative samples) to try to make up for the class imbalance. The ratios chosen were inversely proportional to the prevalence of the class.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2.">Subtask 2</head><p>The drug name extraction task was reformulated as a drug Named Entity Recognition (NER) task. As the setting for this task was multilingual, we focused on two major approaches: building a single multilingual model capable of making predictions on all languages by leveraging knowledge transfer between languages during training, and training language-specific models which are independent from each other. In addition, as the drug name list is dynamic, we experimented with adding drug names from official drug registries such as DrugBank as a gazzetteer. The gazetteer collection is described in Section 3. <ref type="bibr" target="#b3">4</ref>.</p><p>In particular we experimented with the following methods:</p><p>•</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Multilingual model</head><p>For the foundation multilingual model we used a FacebookAI/xlm-roberta-base<ref type="foot" target="#foot_12">33</ref> backbone.</p><p>We experimented with training the original model and also performing domain adaptation. The adapted model was trained on English, Spanish and Italian medical datasets on the Masked Language Modeling objective using the dataset from Section 3.3. In addition, we experimented with a multilingual model pretrained on several NER datasets which has shown good results on similar tasks: numind/NuNER-multilingual-v0.1<ref type="foot" target="#foot_13">34</ref> • Language-specific models</p><p>As a set of language-specific models we focused on michiyasunaga/BioLinkBERT-base<ref type="foot" target="#foot_14">35</ref> for English, PlanTL-GOB-ES/roberta-base-biomedical-clinical-es<ref type="foot" target="#foot_15">36</ref> for Spanish and dbmdz/bert-base-italian-cased 37 for Italian. As the models for English and Spanish were originally pretrained on medical data, we expected them to perform well as is, whilst for the Italian model, we conducted additional domain adaptation. Furthermore, as Italian medical vocabulary is relatively close to Spanish, we conducted language adaptation of the Spanish model on the Italian medical dataset too.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>• Drug Gazetteer</head><p>The gazetteer was applied before model predictions by finding exact match of drug names in the text. Dictionary statistics and description can be found in Section 3.4. As described in Section 3.2, the dataset is highly imbalanced and as a potential solution for boosting the precision of predictions, we experimented with training a binary classification model to identify sentences that contain drug annotations and sort out empty sentences. FacebookAI/xlm-roberta-base after domain adaptation was used t train the classifier model. Furthermore, as many drug names included punctuation marks, we added a post-processing step joining drug names divided by symbols / and +.</p><p>Figure <ref type="figure" target="#fig_0">1</ref> shows the overall architecture of the approach. Depending on the configuration we either include or do not include filtering and dictionary-based annotations.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Experiments &amp; Results</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.1.">Subtask 1</head><p>For the Named Entity Recognition task we employed a standard approach for token classification task. To simplify the setting and to avoid truncation due to limits in the input sequence length, we trained on the split sentences (see Section 3.5). Standard token classification pipeline from Huggingface Transformers was used. For the pretraining, we used a standard masked language modeling pretraining objective. In general, the models achieved better recall than precision. The models that achieved the best results were the models that had domain adaptation for both Spanish and clinical language domains. Despite that DeBERTa is generally a model that achieves better performance than XLM-R, the best model in our experiments was CLIN-X-ES, a XLM-R-based model pretrained on a Spanish clinical corpus. It is notable that DeBERTa-base achieved better performance than DeBERTa-large probably because the dataset was quite small. Using pretraining generally gave small improvements or no improvements, but pertained models converged much faster -on average it took them between 2-6 epochs less to converge. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.2.">Subtask 2</head><p>For the second subtask the setup was quite similar to the first one. During the first set of experiments we compared the multilingual candidate models. The results are reported at Table <ref type="table" target="#tab_5">6</ref>. It could be observed that in-domain pretraining positively influences the overall performance of the model. For each of the languages, the F-score improves by circa 2%.</p><p>As for the monolingual model comparison reported at Table <ref type="table">7</ref>, we can observe that in general monolingual models exhibit slightly better performance for all the target languages except for Italian. Curiously, a Spanish model trained on medical data and fine-tuned on Italian medical data performs better compared to Italian foundation model adapted to the clinical domain.</p><p>Table <ref type="table" target="#tab_6">8</ref> reports results of the submitted systems on the validation dataset. In general, language-specific models show better performance for this task. Filtering consistently increases precision but slightly  The final evaluation on the test set showed that medical XLM-R is a SOTA result for Italian and BioLinkBERT is the best approach for English.</p><p>As for the final submission, the best candidate models remained unchanged. Although on the validation set filtering out empty sentences proved to be a beneficial strategy, on the test set results of the pipeline without filtering are generally better.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.3.">Error Analysis</head><p>After the annotated test data was released, we conducted error analysis and found the following patterns. First, most of the errors related to recall are the cases of drugs that include numbers and special symbols. However, there were cases where the model predicted drugs names that were missing in the annotations (e.g. it was the case with Angiotensin-converting enzyme (ACE) inhibitors). Based on the performance evaluation one can see that adding dictionary-based matching decreases precision of the models. This can be explained by several factors rooting from the nature of the dictionary we used. In the official drug registries there are quite a lot of drugs with ambiguous brand names (e.g. vita -life in Italian), which increases the number of false positive predictions. In addition, there are drug names that are homonymous with laboratory test measures, for instance sodium, vitamin K, etc. Rule-based matching does not rule out such cases. Lastly, official names of the medications include dosages and concentrations, while in the challenge data those were not included in the annotations (e.g. official name lidocain 2%, annotated name lidocain).</p><p>The issue with drug dosage and medication concentration is relevant for model predictions too. While the Italian models rarely included concentrations in the predictions, for English and Spanish this was often the case. Such inconsistencies originate from training data differences. The final models were trained on 2 different datasets (DrugTEMIST, CardioCCC) and those appeared to be annotated differently. While in DrugTEMIST dosages and concentrations were consistently included in the drug span, for CardioCCC this is not the case. Moreover, combined drug names were also annotated inconsistently being either split into 2 drugs, or combined into one. Figure <ref type="figure" target="#fig_1">2</ref> shows an example of drug annotations which do not include the respective dosages as part of the labelled span. The inconsistency of labelling between the train, dev, and test sets is a source of errors for the final trained model.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Conclusion</head><p>In this paper we presented transformer based models for drug and disease named entity recognition in multilingual clinical texts. The experiments for disease NER showed that the domain-adapted and language-specific model CLIN-X-ES scored 0.819 F1 and outperformed DeBERTa-based models. For the drug NER task, the best scores for English and Spanish were achieved using monolingual models BioLinkBERT and RoBerta-base-biomedical-clinical-es respectively. For Italian, in contrast, the multilingual model XLMR_med outperformed the monolingual one by a small margin. Although dictionaries showed contribution to the recall metric, the ambiguities in the drug names that contain some common vocabulary or the common confusion with lab test results cause a significant drop in the precision. We experimented with different approaches to attempt to tackle the label sparsity issueadjusting class weights during training as well as adding a classification step which predicts whether the sentence contains a drug name. The approach using a classifier as a filtering step showed improved performance on the validation set, however, did not work so well on the actual test set. As future work, using hybrid solution with the help of LLMs can improve the system performance and address the issues with the disambiguation of the term usage in context.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: The architecture of the system for named entity recognition. Optional modules are displayed with a dashed border.</figDesc><graphic coords="6,72.00,477.01,451.27,183.57" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: Inconsistent annotation example.</figDesc><graphic coords="9,72.00,240.61,451.28,72.99" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1</head><label>1</label><figDesc>Number of sentences per dataset after pre-processing</figDesc><table><row><cell>Dataset</cell><cell>No. sentences</cell></row><row><cell>DisTEMIST</cell><cell>15 885</cell></row><row><cell>CardioCCC</cell><cell>19 405</cell></row><row><cell>CardioCCC test + Background</cell><cell>197 430</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2</head><label>2</label><figDesc>Domain adaptation dataset statistics in tokens</figDesc><table><row><cell>Data split</cell><cell cols="2">English Spanish</cell><cell>Italian</cell></row><row><cell>Train</cell><cell cols="3">20 198 997 21 218 064 21 402 113</cell></row><row><cell>Dev</cell><cell>1 935 176</cell><cell>1 494 718</cell><cell>2 000 500</cell></row><row><cell>Test</cell><cell>1 387 661</cell><cell>2 660 891</cell><cell>2 271 500</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 3</head><label>3</label><figDesc>Drug dictionary after cleaning</figDesc><table><row><cell cols="4">Drug Dictionary English Spanish Italian</cell></row><row><cell>Drug names</cell><cell>26 843</cell><cell>51151</cell><cell>39985</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 4</head><label>4</label><figDesc>Number of sentences per dataset after pre-processing</figDesc><table><row><cell>Dataset</cell><cell cols="2">Spanish English</cell><cell>Italian</cell><cell>Total</cell></row><row><cell>DrugTEMIST</cell><cell>15 885</cell><cell>16 342</cell><cell>15 913</cell><cell>48 140</cell></row><row><cell>CardioCCC</cell><cell>19 405</cell><cell>18 738</cell><cell>19 396</cell><cell>57 539</cell></row><row><cell cols="5">CardioCCC test + Background 197 430 195 053 198 084 590 567</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_4"><head>Table 5</head><label>5</label><figDesc>Models performance on dev dataset. Pretaining data 1 is proprietary biomedical data, and data 2 is the Custom Medical Text Dataset (see Section 3.2)</figDesc><table><row><cell>Model</cell><cell cols="3">Token Precision Token Recall Token F1</cell></row><row><cell>roberta-base-biomedical-clinical-es</cell><cell>0.744</cell><cell>0.760</cell><cell>0.752</cell></row><row><cell>CLIN-X-ES</cell><cell>0.814</cell><cell>0.823</cell><cell>0.819</cell></row><row><cell>CLIN-X-ES + weighted loss</cell><cell>0.793</cell><cell>0.828</cell><cell>0.810</cell></row><row><cell>Deberta-v3-base</cell><cell>0.786</cell><cell>0.784</cell><cell>0.785</cell></row><row><cell>mDeberta-v3-base</cell><cell>0.778</cell><cell>0.810</cell><cell>0.793</cell></row><row><cell>Deberta-v3-large</cell><cell>0.785</cell><cell>0.755</cell><cell>0.770</cell></row><row><cell>Deberta-v3-base + pretraining data 1</cell><cell>0.768</cell><cell>0.782</cell><cell>0.775</cell></row><row><cell>Deberta-v3-base + pretraining data 2</cell><cell>0.776</cell><cell>0.802</cell><cell>0.789</cell></row><row><cell>CLIN-X-ES + pretraining data 2</cell><cell>0.818</cell><cell>0.816</cell><cell>0.817</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_5"><head>Table 6</head><label>6</label><figDesc>Multilingual models experimentsreduces recall for medical XLM-R model pipelines for all the languages. The result is quite natural as by filtering out extra sentences we are adding more false negative examples to the final prediction. The most significant difference is observed for Italian.</figDesc><table><row><cell>Model</cell><cell cols="3">F1-es F1-it F1-en F1 overall</cell></row><row><cell>NuNER</cell><cell>0.848 0.849</cell><cell>0.852</cell><cell>0.844</cell></row><row><cell>XLMR</cell><cell>0.851 0.860</cell><cell>0.845</cell><cell>0.849</cell></row><row><cell cols="3">XLMR-med 0.871 0.875 0.863</cell><cell>0.873</cell></row><row><cell>Table 7</cell><cell></cell><cell></cell><cell></cell></row><row><cell>Monolingual models experiments</cell><cell></cell><cell></cell><cell></cell></row><row><cell>Model</cell><cell></cell><cell cols="2">Language F1</cell></row><row><cell cols="2">roberta-base-biomedical-clinical-es</cell><cell>es</cell><cell>0.922</cell></row><row><cell>BioLinkBERT</cell><cell></cell><cell>en</cell><cell>0.878</cell></row><row><cell>ClinicalBERT</cell><cell></cell><cell>en</cell><cell>0.863</cell></row><row><cell cols="2">italian-bert-base-cased</cell><cell>it</cell><cell>0.860</cell></row><row><cell cols="2">italian-bert-base-cased_med</cell><cell>it</cell><cell>0.863</cell></row><row><cell cols="2">roberta-base-biomedical-clinical-es</cell><cell>it</cell><cell>0.866</cell></row><row><cell cols="3">roberta-base-biomedical-clinical-es_it it</cell><cell>0.869</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_6"><head>Table 8</head><label>8</label><figDesc>Comparison of monolingual and multilingual pipeline performance on dev dataset.</figDesc><table><row><cell>Model</cell><cell cols="4">Lang Token Precision Token Recall Token F1</cell></row><row><cell>XLMR_med</cell><cell>es</cell><cell>0.843</cell><cell>0.907</cell><cell>0.874</cell></row><row><cell>XLMR_med</cell><cell>en</cell><cell>0.825</cell><cell>0.895</cell><cell>0.859</cell></row><row><cell>XLMR_med</cell><cell>it</cell><cell>0.836</cell><cell>0.892</cell><cell>0.863</cell></row><row><cell>XLMR_med + filtering</cell><cell>es</cell><cell>0.853</cell><cell>0.902</cell><cell>0.877</cell></row><row><cell>XLMR_med + filtering</cell><cell>en</cell><cell>0.838</cell><cell>0.891</cell><cell>0.864</cell></row><row><cell>XLMR_med + filtering</cell><cell>it</cell><cell>0.855</cell><cell>0.889</cell><cell>0.871</cell></row><row><cell>roberta-base-biomedical-clinical-es</cell><cell>es</cell><cell>0.915</cell><cell>0.928</cell><cell>0.922</cell></row><row><cell>BioLinkBERT</cell><cell>en</cell><cell>0.860</cell><cell>0.897</cell><cell>0.878</cell></row><row><cell cols="2">roberta-base-biomedical-clinical-es_it it</cell><cell>0.864</cell><cell>0.874</cell><cell>0.869</cell></row><row><cell>XLMR_med + filtering + dict1</cell><cell>es</cell><cell>0.610</cell><cell>0.909</cell><cell>0.731</cell></row><row><cell>XLMR_med + filtering + dict1</cell><cell>en</cell><cell>0.770</cell><cell>0.911</cell><cell>0.834</cell></row><row><cell>XLMR_med + filtering + dict1</cell><cell>it</cell><cell>0.569</cell><cell>0.863</cell><cell>0.686</cell></row><row><cell>XLMR_med + filtering + dict2</cell><cell>es</cell><cell>0.615</cell><cell>0.911</cell><cell>0.734</cell></row><row><cell>XLMR_med + filtering + dict2</cell><cell>en</cell><cell>0.794</cell><cell>0.911</cell><cell>0.848</cell></row><row><cell>XLMR_med + filtering + dict2</cell><cell>it</cell><cell>0.574</cell><cell>0.866</cell><cell>0.690</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_7"><head>Table 9</head><label>9</label><figDesc>Test set F1 score on the entity level for all submitted systems</figDesc><table><row><cell>Model</cell><cell cols="3">Spanish English Italian</cell></row><row><cell>XLMR</cell><cell>0.912</cell><cell>0.902</cell><cell>0.884</cell></row><row><cell>XLMR_filtering</cell><cell>0.908</cell><cell>0.901</cell><cell>0.881</cell></row><row><cell>Monolingual</cell><cell>0.924</cell><cell>0.922</cell><cell>0.883</cell></row><row><cell>XLMR_filtering_dict1</cell><cell>0.561</cell><cell>0.873</cell><cell>0.684</cell></row><row><cell>XLMR_filtering_dict2</cell><cell>0.822</cell><cell>0.887</cell><cell>0.681</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_0">https://mimic.mit.edu/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_1">https://query.wikidata.org/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_2">https://icd.who.int/es</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="6" xml:id="foot_3">https://www.eciemaps.sanidad.gob.es/browser/metabuscador</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="7" xml:id="foot_4">https://raw.githubusercontent.com/DiseaseOntology/SymptomOntology/main/src/ontology/symp.owl</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="8" xml:id="foot_5">https://emedicine.medscape.com/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="9" xml:id="foot_6">https://medlineplus.gov/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="28" xml:id="foot_7">https://github.com/medspacy/medspacy</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="29" xml:id="foot_8">https://github.com/dhfbk/tint</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="30" xml:id="foot_9">https://github.com/PlanTL-GOB-ES/SPACCC_Sentence-Splitter</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="31" xml:id="foot_10">http://brat.nlplab.org</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="32" xml:id="foot_11">https://huggingface.co/microsoft/mdeberta-v3-base</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="33" xml:id="foot_12">https://huggingface.co/FacebookAI/xlm-roberta-base</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="34" xml:id="foot_13">https://huggingface.co/numind/NuNER-multilingual-v0.1</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="35" xml:id="foot_14">https://huggingface.co/michiyasunaga/BioLinkBERT-base</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="36" xml:id="foot_15">https://huggingface.co/PlanTL-GOB-ES/roberta-base-biomedical-clinical-es</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="37" xml:id="foot_16">https://huggingface.co/dbmdz/bert-base-italian-cased</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgements</head><p>This work was partially supported by the European Union-NextGenerationEU, through the National Recovery and Resilience Plan of the Republic of Bulgaria [Grant Project No. BG-RRP-2.004-0008] and by Horizon Europe research and innovation programme project RES-Q plus [Grant Agreement No. 101057603], funded by the European Union. Views and opinions expressed are however those of the author(s) only and do not necessarily reflect those of the European Union or the Health and Digital Executive Agency. Neither the European Union nor the granting authority can be held responsible for them.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Limitations</head><p>The methods described in this paper were submitted as part of the MultiCardioNER challenge and were validated only on the challenge datasets. Further investigation on different datasets is needed to explore the generalizability of the approach. The specific labeling approach on the datasets impacts the model performance, as in the case of drug NER the inconsistencies of dosage labelling was a source of errors. In different settings, the labelling guidelines used might be different and therefore the presented approach may not perform as well.</p></div>
			</div>

			<div type="annex">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Appendix A Hyperparameters</head><p>For fine-tuning the models we used the following hyperparameters settings:</p><p>• learning rate: We used AdamW <ref type="bibr" target="#b33">[34]</ref> optimizer with between 1-e5 and 5-e5 learning rate.</p><p>• number of epochs: We experimented between 5-20 epochs depending on how long it took the model to converge. • batch size: Initialized to 8 with gradient accumulation steps 1-2 giving an effective batch size of 8-16 due to GPU memory limitations). • learning rate scheduler: linear For pretraining we used the following hyperparameter settings:</p><p>• learning rate: We used AdamW optimizer with 5-e5 learning rate.</p><p>• number of epochs: 3 epochs due to resource limitations. • weight decay: 0.01.</p><p>• batch size: Initialized to 8 due to GPU memory limitations.</p><p>• learning rate scheduler: linear</p></div>			</div>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Overview of MultiCardioNER task at BioASQ 2024 on Medical Speciality and Language Adaptation of Clinical NER Systems for Spanish, English and Italian</title>
		<author>
			<persName><forename type="first">S</forename><surname>Lima-López</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Farré-Maduell</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Rodríguez-Miret</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Rodríguez-Ortega</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Lilli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Lenkowicz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Ceroni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Kossoff</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Shah</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Nentidis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Krithara</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Katsimpras</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Paliouras</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Krallinger</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">CLEF Working Notes</title>
				<editor>
			<persName><forename type="first">G</forename><surname>Faggioli</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Galuščáková</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>García Seco De Herrera</surname></persName>
		</editor>
		<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Overview of BioASQ 2024: The twelfth BioASQ challenge on Large-Scale Biomedical Semantic Indexing and Question Answering</title>
		<author>
			<persName><forename type="first">A</forename><surname>Nentidis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Katsimpras</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Krithara</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Lima-López</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Farré-Maduell</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Krallinger</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Loukachevitch</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Davydova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Tutubalina</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Paliouras</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Experimental IR Meets Multilinguality, Multimodality, and Interaction. Proceedings of the Fifteenth International Conference of the CLEF Association</title>
				<editor>
			<persName><forename type="first">L</forename><surname>Goeuriot</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Mulhem</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><surname>Quénot</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">D</forename><surname>Schwab</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">L</forename><surname>Soulier</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><forename type="middle">Maria</forename><surname>Di Nunzio</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Galuščáková</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>García Seco De Herrera</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><surname>Faggioli</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</editor>
		<meeting><address><addrLine>CLEF</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2024">2024. 2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Biomedical named entity recognition using deep neural networks with contextual information</title>
		<author>
			<persName><forename type="first">H</forename><surname>Cho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Lee</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">BMC bioinformatics</title>
		<imprint>
			<biblScope unit="volume">20</biblScope>
			<biblScope unit="page" from="1" to="11" />
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Aioner: all-in-one scheme-based biomedical named entity recognition using deep learning</title>
		<author>
			<persName><forename type="first">L</forename><surname>Luo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C.-H</forename><surname>Wei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P.-T</forename><surname>Lai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Leaman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Lu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Bioinformatics</title>
		<imprint>
			<biblScope unit="volume">39</biblScope>
			<biblScope unit="page">310</biblScope>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">A dictionary-guided attention network for biomedical named entity recognition in chinese electronic medical records</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Zhu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Zhao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Akhtar</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Expert Systems with Applications</title>
		<imprint>
			<biblScope unit="volume">231</biblScope>
			<biblScope unit="page">120709</biblScope>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Negation-based transfer learning for improving biomedical named entity recognition and relation extraction</title>
		<author>
			<persName><forename type="first">H</forename><surname>Fabregat</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Duque</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Martinez-Romo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Araujo</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Biomedical Informatics</title>
		<imprint>
			<biblScope unit="volume">138</biblScope>
			<biblScope unit="page">104279</biblScope>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Improving biomedical named entity recognition through transfer learning and asymmetric tri-training</title>
		<author>
			<persName><forename type="first">M</forename><surname>Bhattacharya</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Bhat</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Tripathy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Bansal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Choudhary</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Procedia Computer Science</title>
		<imprint>
			<biblScope unit="volume">218</biblScope>
			<biblScope unit="page" from="2723" to="2733" />
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">An effective undersampling method for biomedical named entity recognition using machine learning</title>
		<author>
			<persName><forename type="first">S</forename><surname>Archana</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Prakash</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Evolving Systems</title>
		<imprint>
			<biblScope unit="page" from="1" to="9" />
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">A comparative study for biomedical named entity recognition</title>
		<author>
			<persName><forename type="first">X</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Yang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Guan</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">International Journal of Machine Learning and Cybernetics</title>
		<imprint>
			<biblScope unit="volume">9</biblScope>
			<biblScope unit="page" from="373" to="382" />
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<monogr>
		<author>
			<persName><forename type="first">A</forename><surname>Yazdani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Proios</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Rouhizadeh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Teodoro</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2302.04185</idno>
		<title level="m">Efficient joint learning for clinical named entity recognition and relation extraction using fourier networks: A use case in adverse drug events</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Dictionary-based matching graph network for biomedical named entity recognition</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Lou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Zhu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Tan</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Scientific Reports</title>
		<imprint>
			<biblScope unit="volume">13</biblScope>
			<biblScope unit="page">21667</biblScope>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Bioaug: Conditional generation based data augmentation for low-resource biomedical ner</title>
		<author>
			<persName><forename type="first">S</forename><surname>Ghosh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">U</forename><surname>Tyagi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Kumar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Manocha</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 46th International ACM SIGIR Conference on Research and Development in Information Retrieval</title>
				<meeting>the 46th International ACM SIGIR Conference on Research and Development in Information Retrieval</meeting>
		<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="1853" to="1858" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Data augmentation via context similarity: An application to biomedical named entity recognition</title>
		<author>
			<persName><forename type="first">I</forename><surname>Bartolini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Moscato</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Postiglione</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Sperlì</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Vignali</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Information Systems</title>
		<imprint>
			<biblScope unit="volume">119</biblScope>
			<biblScope unit="page">102291</biblScope>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">A study of deep learning approaches for medication and adverse drug event extraction from clinical text</title>
		<author>
			<persName><forename type="first">Q</forename><surname>Wei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Ji</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Du</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Xiang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Tiryaki</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Zhang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of the American Medical Informatics Association</title>
		<imprint>
			<biblScope unit="volume">27</biblScope>
			<biblScope unit="page" from="13" to="21" />
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Contextualized medication event extraction with striding ner and multi-turn qa</title>
		<author>
			<persName><forename type="first">T</forename><surname>Tsujimura</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Yamada</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Ida</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Miwa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Sasaki</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Biomedical Informatics</title>
		<imprint>
			<biblScope unit="volume">144</biblScope>
			<biblScope unit="page">104416</biblScope>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Recognition of medication information from discharge summaries using ensembles of classifiers</title>
		<author>
			<persName><forename type="first">S</forename><surname>Doan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Collier</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">H</forename><surname>Duy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">M</forename><surname>Phuong</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">BMC medical informatics and decision making</title>
		<imprint>
			<biblScope unit="volume">12</biblScope>
			<biblScope unit="page" from="1" to="10" />
			<date type="published" when="2012">2012</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Recognizing medication related entities in hospital discharge summaries using support vector machine</title>
		<author>
			<persName><forename type="first">S</forename><surname>Doan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Xu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of COLING. International conference on computational linguistics</title>
				<meeting>COLING. International conference on computational linguistics</meeting>
		<imprint>
			<publisher>NIH Public Access</publisher>
			<date type="published" when="2010">2010. 2010</date>
			<biblScope unit="page">259</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Overview of the 2022 n2c2 shared task on contextualized medication event extraction in clinical notes</title>
		<author>
			<persName><forename type="first">D</forename><surname>Mahajan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">J</forename><surname>Liang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C.-H</forename><surname>Tsou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ö</forename><surname>Uzuner</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Biomedical Informatics</title>
		<imprint>
			<biblScope unit="volume">144</biblScope>
			<biblScope unit="page">104432</biblScope>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">Recognition and extraction of named entities in online medical diagnosis data based on a deep neural network</title>
		<author>
			<persName><forename type="first">X</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Zhou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Wang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Visual Communication and Image Representation</title>
		<imprint>
			<biblScope unit="volume">60</biblScope>
			<biblScope unit="page" from="1" to="15" />
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">Comparing transformer-based ner approaches for analysing textual medical diagnoses</title>
		<author>
			<persName><forename type="first">M</forename><surname>Polignano</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>De Gemmis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Semeraro</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">CLEF (Working Notes)</title>
				<imprint>
			<date type="published" when="2021">2021</date>
			<biblScope unit="page" from="818" to="833" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">Iterative annotation of biomedical ner corpora with deep neural networks and knowledge bases</title>
		<author>
			<persName><forename type="first">S</forename><surname>Silvestri</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Gargiulo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Ciampi</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Applied sciences</title>
		<imprint>
			<biblScope unit="volume">12</biblScope>
			<biblScope unit="page">5775</biblScope>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<monogr>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">P</forename><surname>Carrino</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Armengol-Estapé</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Gutiérrez-Fandiño</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Llop-Palao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Pàmies</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Gonzalez-Agirre</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Villegas</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2109.03570</idno>
		<title level="m">Biomedical and clinical language models for spanish: On the benefits of domain-specific pretraining in a mid-resource scenario</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">Biomedical spanish language models for entity recognition and linking at bioasq distemist</title>
		<author>
			<persName><forename type="first">V</forename><surname>Moscato</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Postiglione</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Sperlì</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Working Notes of Conference and Labs of the Evaluation (CLEF) Forum. CEUR Workshop Proceedings</title>
				<imprint>
			<date type="published" when="2022">2022</date>
			<biblScope unit="page" from="315" to="324" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">Hpi-dhc @ bioasq distemist: Spanish biomedical entity linking with pre-trained transformers and cross-lingual candidate retrieval</title>
		<author>
			<persName><forename type="first">F</forename><surname>Borchert</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M.-P</forename><surname>Schapranow</surname></persName>
		</author>
		<ptr target="https://api.semanticscholar.org/CorpusID:251471783" />
	</analytic>
	<monogr>
		<title level="m">Conference and Labs of the Evaluation Forum</title>
				<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<analytic>
		<title level="a" type="main">Sinai at clef 2022: Leveraging biomedical transformers to detect and normalize disease mentions</title>
		<author>
			<persName><forename type="first">M</forename><surname>Chizhikova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Collado-Montañez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>López-Úbeda</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">C</forename><surname>Díaz-Galiano</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">A U</forename><surname>López</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">T M</forename><surname>Valdivia</surname></persName>
		</author>
		<ptr target="https://api.semanticscholar.org/CorpusID:251471813" />
	</analytic>
	<monogr>
		<title level="m">Conference and Labs of the Evaluation Forum</title>
				<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b25">
	<analytic>
		<title level="a" type="main">Clinical named entity recognition and linking using bert in combination with spanish medical embeddings</title>
		<author>
			<persName><forename type="first">J</forename><surname>Reyes-Aguillón</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Del Moral</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Ramos-Flores</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Gómez-Adorno</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Bel-Enguix</surname></persName>
		</author>
		<ptr target="https://api.semanticscholar.org/CorpusID:251471813" />
	</analytic>
	<monogr>
		<title level="m">Conference and Labs of the Evaluation Forum</title>
				<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b26">
	<analytic>
		<title level="a" type="main">Discovering medical procedures in spanish using transformer models with mcrf and augmentation</title>
		<author>
			<persName><forename type="first">T</forename><surname>Almeida</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">A A</forename><surname>Jonker</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Poudel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">M</forename><surname>Silva</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Matos</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Working Notes of CLEF 2023 -Conference and Labs of the Evaluation Forum</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b27">
	<analytic>
		<title level="a" type="main">Coming a long way with pre-trained transformers and string matching techniques: Clinical procedure mention recognition and normalization</title>
		<author>
			<persName><forename type="first">M</forename><surname>Chizhikova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Collado-Montañez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">C</forename><surname>Díaz-Galiano</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">A</forename><surname>Ureña-López</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">T</forename><surname>Martín-Valdivia</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Working Notes of CLEF 2023 -Conference and Labs of the Evaluation Forum</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b28">
	<analytic>
		<title level="a" type="main">ICB-UMA at BioCreative VIII @ AMIA 2023 Task 2 SYMPTEMIST (Symptom TExt Mining Shared Task)</title>
		<author>
			<persName><forename type="first">F</forename><surname>Gallego</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">J</forename><surname>Veredas</surname></persName>
		</author>
		<idno type="DOI">10.5281/zenodo.10104058</idno>
		<ptr target="https://doi.org/10.5281/zenodo.10104058.doi:10.5281/zenodo.10104058" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the BioCreative VIII Challenge and Workshop: Curation and Evaluation in the era of Generative Models</title>
				<meeting>the BioCreative VIII Challenge and Workshop: Curation and Evaluation in the era of Generative Models<address><addrLine>Zenodo</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b29">
	<monogr>
		<author>
			<persName><forename type="first">A</forename><surname>Palmero Aprosio</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Moretti</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1609.06204</idno>
		<title level="m">Italy goes to Stanford: a collection of CoreNLP modules for Italian</title>
				<imprint>
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
	<note type="report_type">ArXiv e-prints</note>
</biblStruct>

<biblStruct xml:id="b30">
	<analytic>
		<title level="a" type="main">brat: a web-based tool for NLP-assisted text annotation</title>
		<author>
			<persName><forename type="first">P</forename><surname>Stenetorp</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Pyysalo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Topić</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Ohta</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Ananiadou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Tsujii</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Demonstrations Session at EACL 2012, Association for Computational Linguistics</title>
				<meeting>the Demonstrations Session at EACL 2012, Association for Computational Linguistics<address><addrLine>Avignon, France</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2012">2012</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b31">
	<monogr>
		<author>
			<persName><forename type="first">L</forename><surname>Lange</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Adel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Strötgen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Klakow</surname></persName>
		</author>
		<ptr target="https://arxiv.org/abs/2112.08754.arXiv:2112.08754" />
		<title level="m">Clin-x: pre-trained language models and a study on cross-task transfer for concept extraction in the clinical domain</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b32">
	<monogr>
		<author>
			<persName><forename type="first">P</forename><surname>He</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Gao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Chen</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2111.09543</idno>
		<title level="m">Debertav3: Improving deberta using electra-style pre-training with gradientdisentangled embedding sharing</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b33">
	<monogr>
		<title level="m" type="main">Decoupled weight decay regularization</title>
		<author>
			<persName><forename type="first">I</forename><surname>Loshchilov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Hutter</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1711.05101</idno>
		<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
