<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Natural Language Processing Models for Knowledge Discovery in Medical Texts</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author role="corresp">
							<persName><forename type="first">Eduardo</forename><surname>Grande</surname></persName>
							<email>eduardo.grande@ua.es</email>
							<affiliation key="aff0">
								<orgName type="department">Department of Software and Computing Systems</orgName>
								<orgName type="institution">University of Alicante</orgName>
								<address>
									<country key="ES">Spain</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Natural Language Processing Models for Knowledge Discovery in Medical Texts</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">29C433D2A849924C4867AEB42FBBFA2E</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T19:02+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Natural Language Processing</term>
					<term>Knowledge Discovery</term>
					<term>Large Language Models</term>
					<term>Electronic Health Records</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Electronic Health Records (EHRs) contain vast amounts of valuable information about patients' diseases, diagnoses, or medications, mostly in an unstructured format. Recently, Large Language Models (LLMs), particularly generative models, have gained popularity due to their remarkable capabilities. This PhD thesis aims to harness the power of these models for the medical field, specifically for knowledge extraction tasks. The goal is to adjust NLP models to extract critical insights from EHRs and other medical texts. However, one of the main challenges is the limited availability of publicly accessible medical data, especially annotated datasets in languages other than English. In order to adjust the models, the thesis explores various adaptation techniques, including prompt-tuning or continual pre-training to enhance the models' ability to process medical information effectively. Additionally, it evaluates different LLM architectures to determine the most suitable for medical knowledge extraction. Innovative strategies like LoRA (Low-Rank Adaptation) and QLoRA (Quantized LoRA) are also investigated to try to improve efficiency. The outcomes of this research hold the potential to significantly enhance healthcare delivery and and help practitioners quickly understand patient data.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Electronic Health Records (EHRs) are text documents that contain all medical information about patients. They maintain extensive records regarding patients' health, including biographical information, disease history, symptoms, diagnoses, medication prescriptions, and other relevant patient information.</p><p>The information contained in EHRs is often unstructured and written in free form, meaning there are no standardized formats for recording patient data. Only a few documents or pieces of information, such as lab tests that include blood measurements, can be considered structured. This lack of structure in the majority of documents makes it challenging to detect, extract, and classify the highly valuable information they contain.</p><p>The aim of the present doctoral work is to create Natural Language Processing (NLP) models that can automatically extract knowledge, not only from EHRs but also from other unstructured medical texts, by training them using collected publicly available corpora. The extracted knowledge can be extremely valuable for predicting patients' diseases, conducting population-based public health studies, or analyzing the impact of a disease on a population group.</p><p>Creating such models requires annotated data. In the medical field, the scarcity of these resources is a significant challenge. Most published works use private datasets obtained through specific agreements between researchers and medical institutions.</p><p>As explained by Li et al. <ref type="bibr" target="#b0">[1]</ref>, most works use private datasets, and those using public ones mainly rely on the MIMIC datasets <ref type="bibr" target="#b1">[2,</ref><ref type="bibr" target="#b2">3,</ref><ref type="bibr" target="#b3">4]</ref> or the i2b2/n2c2 datasets. Both datasets contain annotated health documents in English that have been extensively used for developing medical NLP models.</p><p>When seeking data in other languages, such as Spanish, the difficulty increases significantly. A good example demonstrating this scarcity is the work done by Carrino et al. <ref type="bibr" target="#b4">[5]</ref>, which aimed to create a biomedical and medical language model for Spanish. The data they used primarily consisted of scientific publications, patents, or Wikipedia articles rather than EHRs. The only corpus composed of medical cases was not publicly available.</p><p>The strategy of using a mixture of documents, rather than exclusively using medical cases for training NLP models, is common. Many recent works resort to this approach to mitigate the scarcity of valuable data, such as EHRs. These documents often include publications available in PubMed, medical-related documents crawled from the internet, or even the UMLS ontology, as explained by Wornow et al. <ref type="bibr" target="#b5">[6]</ref>.</p><p>Currently, NLP researchers predominantly use Large Language Models (LLMs) to solve common field tasks. As shown in various reviews <ref type="bibr" target="#b0">[1,</ref><ref type="bibr" target="#b5">6,</ref><ref type="bibr" target="#b6">7]</ref>, most research works focus on using BERT models, or regarding the newest LLMs, most of them use GPT models (via the OpenAI API).</p><p>To date, almost no work has been done using the newest LLM models, such as LLaMa-3 or Mistral. Exploring these newly released models could yield promising results for knowledge extraction tasks.</p><p>In the following sections, the details of the proposal for this doctoral work will be described.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Background and related work</head><p>In the last two years, several surveys have been published concerning the use of LLMs in the medical field.</p><p>• Li et al. <ref type="bibr" target="#b0">[1]</ref> presented a survey based on 329 related papers, explaining the evolution of LLMs, their different architecture types, what an EHR is, publicly available EHR datasets, methods of fine-tuning the models, and research trends in different NLP tasks, such as Named Entity Recognition (NER) and Information Extraction. • Wornow et al. <ref type="bibr" target="#b5">[6]</ref> conducted a survey about the use of Functional Models (FMs-models capable of performing many different tasks) in the medical field. They explained the different FMs specifically created for medical purposes (medical models, EHR models, etc.), their benefits, available public medical data, how the models have been evaluated, and future trends. • Huang et al. <ref type="bibr" target="#b6">[7]</ref> provided a comprehensive survey presenting various aspects of using LLMs in the medical field. They showed the different applications for which LLMs have been used, such as data processing (including the NER task) and various models. They also extensively explained the metrics and benchmarks used to measure the performance of these models for specific tasks.</p><p>Regarding related work, several specific relevant works are highlighted for this research:</p><p>• Guevara et al. <ref type="bibr" target="#b7">[8]</ref> used LLMs to extract Social Determinants of Health (SDoH), which are conditions surrounding patients that affect their health. Their work is particularly interesting because they employed different approaches and models to achieve the goal of extracting SDoH from EHRs. They also used techniques such as LoRA <ref type="bibr" target="#b8">[9]</ref> and PEFT <ref type="bibr" target="#b9">[10]</ref> for efficiently adjusting the models. • de la Iglesia et al. <ref type="bibr" target="#b10">[11]</ref> created a corpus composed of 1038 Electronic Clinical Narratives (ECNs) written in Spanish. It contains annotations of seven different types, referring to illnesses, medications, or treatments. This corpus could be used as a base to adjust and test an LLM for detecting information and classifying it into different categories. • Ahsan et al. <ref type="bibr" target="#b11">[12]</ref> extracted evidence from EHRs using LLMs. By prompting models like Flan-T5 XXL or Mistral multiple times, they obtained evidence and corresponding explanations from the models, where the prompts contained information from given EHRs. This approach of directly prompting LLMs without fine-tuning could be explored as a baseline to compare with results obtained from fine-tuned models. • Agrawal et al. <ref type="bibr" target="#b12">[13]</ref> aimed to extract information from medical texts without fine-tuning models.</p><p>They used GPT-3 for information extraction, focusing on prompting the model and parsing the output to get structured results, such as arrays of strings. Parsing the results of generative models is crucial for extracting complex information accurately.</p><p>• Gallego et al. <ref type="bibr" target="#b13">[14]</ref> presented a pipeline for medical entity linking, emphasizing the use of standardized terminologies like UMLS or SNOMED-CT. These ontologies could be considered when performing knowledge extraction from medical documents, as linking information with the corresponding codes would normalize the data, facilitating integration with information retrieved by other systems.</p><p>These works have been instrumental in understanding the current trends in the field of knowledge extraction from medical texts. Given the recent popularity of LLMs, most of the surveys are quite recent, indicating that this research field is currently booming.</p><p>In the following sections, the proposed research and planned experiments will be described. These plans have been formulated after reviewing the mentioned surveys and similar experiments conducted by other researchers.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Description of the proposed research</head><p>This work plans to use LLMs for knowledge extraction, primarily exploring the NER task, in the medical field. Figure <ref type="figure" target="#fig_0">1</ref> illustrates how field-related data, such as EHRs, and knowledge bases, such as SNOMED-CT or UMLS, can be used to train an LLM for extracting knowledge from plain text. This enables the generation of structured information, such as text annotations for disease or pharmacological substance mentions. As explained previously, almost no work has explored the use of currently published LLMs such as LLaMA-3, Mistral, or Claude 3. The main research will be exploring how to adapt these models to the knowledge extraction task. As outlined in the justification section, the majority of the available data is in English, with a limited number of corpora in Spanish. Therefore, the languages used for data extraction will primarily depend on the availability of data.</p><p>Many adaptation techniques can be explored, but the main one to explore is prompt-tuning or adapter-tuning. A set of templates can be defined, which can be applied over several corpus, similar to how Google does with FLAN <ref type="bibr" target="#b14">[15]</ref>. By doing this, the models may learn specifically the task we want them to, transitioning from a generalist position towards a specific one.</p><p>Not only these adapting techniques are going to be explored. Continual pre-training techniques could also be explored. Several plain text datasets from the medical field could be collected and then continue the training of an LLM to incorporate new target vocabulary.</p><p>Regarding the data needed to undertake the work, we can distinguish between the necessity of data for continual pre-training (this process just needs plain text data) and the data for adjusting the models (annotated data).</p><p>For the non-annotated data, an exploration of available datasets and corpus can be done. Collecting new not-annotated data is easier than the annotated one, so collecting new data by crawling websites can be done, always taking in mind the rights and licenses of the crawled pages to avoid violating any rules or applicable laws.</p><p>On the other side, for the annotated data needed, in a first step the public available datasets are going to be explored. There's a need to explore if their data is within the desired scope and if their annotations are significant. Foreseeably, these data will be scarce, so more data will have to be collected. If more health records want to be obtained, alliances between medical institutions such as public hospitals could be arranged. Other sources of these data could be asking researches who have already use datasets of EHRs for sharing them under license and terms of use.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Methodology and Proposed Experiments</head><p>The methodology to achieve the established goals is based on achieving milestones planned for the three years of the doctoral research.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1.">Year 1</head><p>• State-of-the-art Study: Begin with an extensive review of current techniques used for creating LLMs specialized in knowledge extraction. • Initial Experiments: Start experimenting with some of these LLMs. This involves identifying and obtaining publicly available medical data to apply and test these models.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2.">Year 2</head><p>• Continued Research on LLMs: Deepen the research on knowledge extraction LLMs and continue searching for medical data. This may involve forming agreements with hospitals or other healthcare institutions to access non-public datasets. • Data Storage and Representation: Once the extracted knowledge is stored, representation methods should be researched to illustrate the knowledge extracted by the created models.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.3.">Year 3</head><p>• Completion of Research: Finalize the research activities initiated in the previous years.</p><p>• Publications and Thesis Writing: Complete pending scientific publications and start writing the doctoral thesis.</p><p>Regarding the proposed experiments, they will evolve based on the results obtained from initial trials. Some of the planned experiments include:</p><p>Prompt-Tuning LLM: Choose an LLM and perform prompt-tuning. This involves exploring different tuning methods, optimal hyperparameters, optimizers, and evaluation metrics.</p><p>In-Context Learning: Apply the prompt-tuning technique to an LLM. This involves creating and testing various sets of prompts. In-context learning uses textual inputs (prompts) to fine-tune an LLM. Different prompt construction strategies, such as one-shot, few-shot, and chain-of-thought, will be tested.</p><p>LoRA and QLoRA: Test the LoRA (Low-Rank Adaptation) and QLoRA (Quantized Low-Rank Adaptation) techniques for prompt-tuning. LoRA <ref type="bibr" target="#b8">[9]</ref> involves training only a reduced set of model parameters, specifically low-rank representations, to reduce training time and GPU usage. QLoRA <ref type="bibr" target="#b15">[16]</ref> extends LoRA by representing model weights in 4-bit precision, further reducing memory usage and improving efficiency.</p><p>Continual Pre-training: Investigate how the performance of an LLM improves with increased exposure to medical vocabulary. Conduct continual pre-training of an LLM with medical-specific vocabulary to enhance its performance in knowledge extraction tasks.</p><p>Use of Synthetic Generated Data: Address the scarcity of publicly available medical datasets by generating synthetic annotated data. Combine synthetic data with real data to increase the number of training examples, potentially improving model performance.</p><p>As the research progresses and new state-of-the-art publications are reviewed, additional experiments will be conceived. These may involve employing new techniques, testing new models, and exploring different methods for model adjustment. This dynamic approach ensures the research remains at the cutting edge of advancements in NLP and medical data extraction.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Specific Issues of Research to be Discussed</head><p>In this section, and once the doctoral work has been explained, various questions are posed for further discussion.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Q1. Sources of medical data</head><p>To achieve the objectives of the present project, medical data is needed.</p><p>Across the document, several ways of obtaining these data have been explained. Are these ways of obtaining the data useful? Could other ways be explored? Are there public medical datasets available that have not been mentioned? The resultant discussion produced by the presented questions, as well as other aspects that could arise, would be enriching for the PhD thesis.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Overview of the main goal of the doctoral work</figDesc><graphic coords="3,72.00,348.90,451.26,182.42" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Q2.</head><label></label><figDesc>Which LLM architecture is better to use? In recent years, most state-of-the-art LLMs have a decoder-only architecture, while the most-used LLM in the medical field has been BERT. Which architecture is better for knowledge extraction in the medical field? How much can the architecture of a model influence the resultant performance? Q3. What is the better way of adjusting the LLMs to the knowledge extraction task? Throughout the work, different ways of adjusting models have been presented, with prompt-tuning being the most common. What is the best way of performing prompt-tuning on an LLM? Are there any other techniques besides LoRA and QLoRA?</figDesc></figure>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Acknowledgements</head><p>The publication is part of the grant PRE2022-101573, funded by MCIN/AEI/10.13039/501100011033 and the ESF+.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<author>
			<persName><forename type="first">L</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Zhou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Gao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Hua</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Fan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Yu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Hagen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">L</forename><surname>Assimes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Hemphill</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2405.03066</idno>
		<title level="m">A scoping review of using large language models (llms) to investigate electronic health records (ehrs)</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Mimic ii: a massive temporal icu patient database to support research in intelligent patient monitoring</title>
		<author>
			<persName><forename type="first">M</forename><surname>Saeed</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Lieu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Raber</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">G</forename><surname>Mark</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Computers in cardiology</title>
				<imprint>
			<publisher>IEEE</publisher>
			<date type="published" when="2002">2002</date>
			<biblScope unit="page" from="641" to="644" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Mimic-iii, a freely accessible critical care database</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">E</forename><surname>Johnson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">J</forename><surname>Pollard</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Shen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L.-W</forename><forename type="middle">H</forename><surname>Lehman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Feng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Ghassemi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Moody</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Szolovits</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Anthony Celi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">G</forename><surname>Mark</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Scientific data</title>
		<imprint>
			<biblScope unit="volume">3</biblScope>
			<biblScope unit="page" from="1" to="9" />
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Mimic-iv, a freely accessible electronic health record dataset</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">E</forename><surname>Johnson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Bulgarelli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Shen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Gayles</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Shammout</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Horng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">J</forename><surname>Pollard</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Hao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Moody</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Gow</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Scientific data</title>
		<imprint>
			<biblScope unit="volume">10</biblScope>
			<biblScope unit="page">1</biblScope>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Pretrained biomedical language models for clinical NLP in Spanish</title>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">P</forename><surname>Carrino</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Llop</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Pàmies</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Gutiérrez-Fandiño</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Armengol-Estapé</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Silveira-Ocampo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Valencia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Gonzalez-Agirre</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Villegas</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/2022.bionlp-1.19</idno>
		<ptr target="https://aclanthology.org/2022.bionlp-1.19.doi:10.18653/v1/2022.bionlp-1.19" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 21st Workshop on Biomedical Language Processing, Association for Computational Linguistics</title>
				<meeting>the 21st Workshop on Biomedical Language Processing, Association for Computational Linguistics<address><addrLine>Dublin, Ireland</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2022">2022</date>
			<biblScope unit="page" from="193" to="199" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<monogr>
		<author>
			<persName><forename type="first">M</forename><surname>Wornow</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Thapa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Patel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Steinberg</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Fleming</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">A</forename><surname>Pfeffer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Fries</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><forename type="middle">H</forename><surname>Shah</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2303.12961</idno>
		<title level="m">The shaky foundations of clinical foundation models: a survey of large language models and foundation models for emrs</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b6">
	<monogr>
		<author>
			<persName><forename type="first">Y</forename><surname>Huang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Tang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Chen</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2404.15777</idno>
		<title level="m">A comprehensive survey on evaluating large language model applications in the medical industry</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Large language models to identify social determinants of health in electronic health records</title>
		<author>
			<persName><forename type="first">M</forename><surname>Guevara</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Thomas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">L</forename><surname>Chaunzwa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Franco</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><forename type="middle">H</forename><surname>Kann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Moningi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">M</forename><surname>Qian</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Goldstein</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Harper</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">NPJ digital medicine</title>
		<imprint>
			<biblScope unit="volume">7</biblScope>
			<biblScope unit="page">6</biblScope>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<monogr>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">J</forename><surname>Hu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Shen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Wallis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Allen-Zhu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Chen</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2106.09685</idno>
		<title level="m">Lora: Low-rank adaptation of large language models</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b9">
	<monogr>
		<author>
			<persName><forename type="first">S</forename><surname>Mangrulkar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Gugger</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Debut</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Belkada</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Paul</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Bossan</surname></persName>
		</author>
		<ptr target="https://github.com/huggingface/peft" />
		<title level="m">Peft: State-of-the-art parameterefficient fine-tuning methods</title>
				<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">An open source corpus and automatic tool for section identification in spanish health records</title>
		<author>
			<persName><forename type="first">I</forename><surname>De La Iglesia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Vivó</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Chocrón</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>De Maeztu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Gojenola</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Atutxa</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Biomedical Informatics</title>
		<imprint>
			<biblScope unit="volume">145</biblScope>
			<biblScope unit="page">104461</biblScope>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<monogr>
		<author>
			<persName><forename type="first">H</forename><surname>Ahsan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">J</forename><surname>Mcinerney</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Kim</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Potter</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Young</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Amir</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><forename type="middle">C</forename><surname>Wallace</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2309.04550</idno>
		<title level="m">Retrieving evidence from ehrs with llms: Possibilities and challenges</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b12">
	<monogr>
		<author>
			<persName><forename type="first">M</forename><surname>Agrawal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Hegselmann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Lang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Kim</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Sontag</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2205.12689</idno>
		<title level="m">Large language models are few-shot clinical information extractors</title>
				<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b13">
	<monogr>
		<author>
			<persName><forename type="first">F</forename><surname>Gallego</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>López-García</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Gasco-Sánchez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Krallinger</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">J</forename><surname>Veredas</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2404.06367</idno>
		<title level="m">Clinlinker: Medical entity linking of clinical concept mentions in spanish</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b14">
	<monogr>
		<author>
			<persName><forename type="first">J</forename><surname>Wei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Bosma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><forename type="middle">Y</forename><surname>Zhao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Guu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">W</forename><surname>Yu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Lester</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Du</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">M</forename><surname>Dai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><forename type="middle">V</forename><surname>Le</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2109.01652</idno>
		<title level="m">Finetuned language models are zero-shot learners</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Qlora: Efficient finetuning of quantized llms</title>
		<author>
			<persName><forename type="first">T</forename><surname>Dettmers</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Pagnoni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Holtzman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Zettlemoyer</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Advances in Neural Information Processing Systems</title>
		<imprint>
			<biblScope unit="volume">36</biblScope>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
