<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Improving the accessibility of EU laws: the Chat-EUR-Lex project</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Manola</forename><surname>Cherubini</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Institute of Legal Informatics and Judicial Systems (IGSG-CNR)</orgName>
								<address>
									<addrLine>via dei Barucci 20</addrLine>
									<postCode>50127</postCode>
									<settlement>Florence</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Francesco</forename><surname>Romano</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Institute of Legal Informatics and Judicial Systems (IGSG-CNR)</orgName>
								<address>
									<addrLine>via dei Barucci 20</addrLine>
									<postCode>50127</postCode>
									<settlement>Florence</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
							<affiliation key="aff1">
								<orgName type="department">Aptus.AI</orgName>
								<address>
									<addrLine>Largo Padre Renzo Spadoni 1</addrLine>
									<postCode>56126</postCode>
									<settlement>Pisa</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Andrea</forename><surname>Bolioli</surname></persName>
							<affiliation key="aff1">
								<orgName type="department">Aptus.AI</orgName>
								<address>
									<addrLine>Largo Padre Renzo Spadoni 1</addrLine>
									<postCode>56126</postCode>
									<settlement>Pisa</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Lorenzo</forename><surname>De Mattei</surname></persName>
						</author>
						<author>
							<persName><forename type="first">Mattia</forename><surname>Sangermano</surname></persName>
							<affiliation key="aff1">
								<orgName type="department">Aptus.AI</orgName>
								<address>
									<addrLine>Largo Padre Renzo Spadoni 1</addrLine>
									<postCode>56126</postCode>
									<settlement>Pisa</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Improving the accessibility of EU laws: the Chat-EUR-Lex project</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">198371CD186F53BF6E0016B3D1C36613</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T16:54+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Legal Informatics</term>
					<term>Large Language Models (LLMs)</term>
					<term>Retrieval Augmented Generation (RAG) 1</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>In this article we describe the results of an ongoing research project on the use of Chat-Based Large Language Models (Chat LLMs) and Retrieval Augmented Generation (RAG) for the access to legal repositories. We are integrating Chat LLMs and RAG to access a dataset of legal acts in English and Italian (a subset of EUR-Lex collection), and interact through a chatbot. We present the state of the art, the objectives, the use cases, the methodology used in the project, and then we discuss the preliminary results.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>In this article we describe the partial results of an ongoing research project on the use of Chat-Based Large Language Models (Chat LLMs) and Retrieval Augmented Generation (RAG) for the access to normative repositories.</p><p>In the project, we are integrating Chat LLMs and RAG to access a dataset of legal documents (European legal acts taken from EUR-Lex repository) and to allow the user to interact through a chatbot.</p><p>In the next sections, we will present the state of the art (2. Related works), the objectives and the methodology used (3. Chat-EUR-Lex methodology), the results of a research survey (4. Research survey), the system architecture (5. System architecture), and then we discuss the results presented in the previous sections (6. Discussion and conclusions). † These authors contributed equally. manola.cherubini@igsg.cnr.it; francesco.romano@igsg.cnr.it; andrea.bolioli@aptus.ai; lorenzo@aptus.ai; mattia.sangermano@aptus.ai</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Related works</head><p>As stated in <ref type="bibr" target="#b0">[1]</ref> and many other sources, "Legal professionals rely on accurate and up-to-date information to make informed decisions, interpret laws, and provide legal counsel". The phenomenon of hallucination and nonsensical outputs of systems based on LLMs is obviously not acceptable in the legal context. To the best of our knowledge, the first survey on the challenges faced by LLMs in the legal domain was presented in <ref type="bibr" target="#b1">[2]</ref>, but mainly for Chinese language. While in other domains, such the financial one, a few LLMs have already been developed <ref type="bibr" target="#b2">[3]</ref>. Large language models are also used in healthcare where LLMs are useful for processing and understanding medical text data, providing valuable insights, and supporting clinical decision-making <ref type="bibr" target="#b3">[4]</ref>.</p><p>LLMs are posing interesting challenges to those who are experimenting with these technologies in the legal field, where the "complexities of legal language, nuanced interpretations, and the ever-evolving nature of legislation present unique challenges that 0000-0002-0242-6633 (Manola Cherubini); 0000-0001-5250-7733 (Francesco Romano); 0000-0003-1681-9435 (Andrea Bolioli) require tailored solutions" <ref type="bibr" target="#b0">[1]</ref>. There are many questions and fears about the actual use of these artificial intelligence tools, e.g., their opacity <ref type="bibr" target="#b4">[5]</ref> and the possibility of hallucinations, but also "legal problems concerning intellectual property, data privacy, and bias and discrimination" <ref type="bibr" target="#b5">[6]</ref>. For this reason, in the European Union it has been decided to regulate the use of artificial intelligence in specific sectors, but also to adopt a regulation that provides for a regulatory framework of reference only for highrisk AI systems <ref type="bibr" target="#b6">[7]</ref>. Some experiments conducted on legal datasets show that LLMs can improve the performance of document page classification <ref type="bibr" target="#b7">[8]</ref>[9], the annotation of legal texts <ref type="bibr" target="#b9">[10]</ref>, the summarization of legal texts <ref type="bibr" target="#b10">[11]</ref>  <ref type="bibr" target="#b11">[12]</ref>, the legal rule classification <ref type="bibr" target="#b12">[13]</ref>, the legal statute identification from facts <ref type="bibr" target="#b13">[14]</ref> and the mining of legal argument <ref type="bibr" target="#b14">[15]</ref>. Other trials explore the ability of LLMs "to explain legal concepts from statutory provisions to legal professionals" <ref type="bibr" target="#b15">[16]</ref> and to create "a register of obligations from various types of legislative and regulatory material" <ref type="bibr" target="#b16">[17]</ref>.</p><p>Other uses can also be mentioned, such as LLMs as legal tutors in the context of legal training <ref type="bibr" target="#b17">[18]</ref> and in one of the most basic tasks required of lawyers, the so-called "statutory reasoning" <ref type="bibr" target="#b18">[19]</ref>. Recently, legislative drafting experiments have been carried out with ChatGPT, particularly for "the comparison of legislation among jurisdictions and the synthesis of the best possible policy for the country based on this comparison" <ref type="bibr" target="#b19">[20]</ref>.</p><p>As it is known, generative AI models have been found to hallucinate, i.e., they can generate false or nonsensical statements <ref type="bibr" target="#b20">[21]</ref>  <ref type="bibr" target="#b21">[22]</ref>., two strategies for reducing this problem are Fine-tuning and Retrieval-Augmented Generation (RAG) <ref type="bibr" target="#b22">[23]</ref>. In both cases, we try to provide the LLM with the relevant information (according to a domain or a specific query). Finetuning involves additional training on a specific dataset, tailoring the model to certain tasks or domains <ref type="bibr">[24] [25]</ref>. This improves accuracy but limits the model to knowledge up to the last fine-tuning. RAG merges a pre-trained model with a retrieval system, accessing current data for accurate responses on recent or specific topics. Its success hinges on the quality of retrieved information and requires maintaining a large, updated database. Both methods enhance model performance in specific areas, balancing current information and resource needs.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Chat-EUR-Lex metodology</head><p>In this section, we present the main problems faced and the methodologies used in the ongoing Chat-EUR-Lex project. Our objective is to create an AI-powered conversational interface that deals with complex legal texts (the regulations in English and Italian published in the EUR-Lex repository), can provide simplified explanations, and allows the user to conduct contextspecific interaction. To present the methodologies used, we describe the activities performed:</p><p>• Legal and Ethics risk assessment. We performed a legal and ethics assessment. The prototype will be compliant to GDPR and EU AI Act, i.e., we will comply with the rules set by EU regulations.</p><p>• UX Research and Survey. We collected data and information from a sample of potential users through a questionnaire, both in Italian and in English. Objective of the survey: understand the needs of people using digital legal resources, and their level of satisfaction; identify users' needs and desires regarding chatbot interaction; know the fears related to the use of generative AI. User experience (UX) research involves studying how users interact with the current EUR-Lex system and identifying pain points and challenges.</p><p>• Data collection. Chat-EUR-Lex dataset comprises a selection of in force legal acts in English and Italian sourced from EUR-Lex, covering the period from January 1, 2014, to December 31, 2023. Specifically, it includes all historical texts preserved in Celex 3 sector that remain unaltered over time, along with the most recent consolidated versions in Celex 0 sector for acts that have undergone amendments. Corrigenda are omitted from this dataset. Additionally, the EUR-Lex documents that are not provided with XML or HTML data are excluded from the selection. Number of documents in English: 19062; documents in Italian: 18164.</p><p>• Semantic search engine setup. Semantic search must allow users to find relevant legal information even if they don't use precise legal terminology. This involves using Natural Language Processing (NLP) techniques, particularly neural embedding, such as the one introduced by (Lai et al.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>2023).</head><p>• RAG-based Chat system development. RAG combines retrieval-based methods with generative language models to provide accurate and contextually relevant responses to user queries. The user can read both the generated answer and the relevant sources, i.e. the portions of regulations used to generate the answer.</p><p>• First version release (June 2024). The first version of the prototype is released to a selected group of users. This version should provide basic functionality and serve as a starting point for further improvements.</p><p>• Feedback collection and tuning. User feedback is actively collected and analysed. This feedback is used to identify areas for improvement and fine-tune both the chat system and the user interface. This iterative process continues to enhance the system's effectiveness and user satisfaction.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Research survey</head><p>In this section we present the results of the questionnaire distributed from December 28, 2023, to March 31, 2024, aimed at legal professionals, law researchers, public officials in the legal sector, compliance specialists, and other people interested in the use of Generative AI in the legal domain, in Italy and other European countries. The objectives of the questionnaire were to understand the needs of people using digital legal resources (EUR-Lex in particular) and their level of satisfaction; identify users' needs and desires regarding chatbot interaction; know the fears related to the use of generative AI. The questionnaire was anonymous; the languages used were Italian and English. We distributed it online on websites and with targeted e-mail activity.</p><p>The We report here some statistics on the Italian responses: 54% of the respondents are legal experts (law researchers, jurists, lawyers, compliance specialists, etc.), while 46% are not legal experts. 66% say that they consulted the EUR-Lex repository at least once. When asked which tool they mainly use to search for legal documents and regulations, 48% answer mainly Google search, 37% mainly EUR-Lex search engine, 15% mainly other tools (we do not report here the answers on the other legal sources). 60.4% say that a generative AI chatbot could help search and interaction, 33.3% don't know, 6.2% say No. The question "Do you know what generative AI is and/or are you a user of generative AI tools?" is answered: 51% "Little", 27% say "Yes, I use them regularly", 22% "Not at all" (remember that these percentages concern responses in Italy). Finally, 87% think generative AI must be regulated, 8% don't know, 5% answer No. For reasons of space, we do not report here the answers to the question "What kind of requests would you make to the chatbot?".</p><p>In summary, these responses allow us to assess the level of knowledge of legal experts in generative artificial intelligence, to see if there are differences between legal experts and other people, to know their fears on these issues, and, above all, to collect the needs and requirements of potential users of the chatbot.</p><p>A detailed report containing the complete questionnaire, the aggregate results and a detailed analysis will be published in May 2024 on the GitHub project repository.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">System architecture</head><p>The pipeline of Chat-EUR-Lex prototype is divided into main parts:</p><p>• An asynchronous batched pipeline which collects and indexes the documents from EUR-Lex into a search engine.</p><p>• A synchronous pipeline that gets the users' queries, retrieves relevant contextual information and provides a response to the users. The asynchronous batched pipeline comprises three main components:</p><p>1. A crawler that collects the data from EUR-Lex. 2. A chunker who chunks the documents into smaller segments.</p><p>3. An embedding model that transforms the segments into dense vectors to be indexed in the vector DB. The synchronous pipeline comprises two main components:</p><p>• A retriever that transforms the query into a vector using the same embedding models used by the asynchronous pipeline and looks into the vector DB for similar contents.</p><p>• An LLM that gets both the query and the context inserted in a prompt template and produces a response to be provided to the users Each time the user does a new query, the whole chat history is passed to the LLM until the maximum prompt length is reached; in that case, older chat parts are truncated.</p><p>This process involves several parameters to be selected, such as:</p><p>• Chunking techniques and size.</p><p>• Embedding models.</p><p>• K-nearest neighbor search techniques.</p><p>• Prompt templates.</p><p>• LLM and its parameters.</p><p>In summary, the RAG approach is a blend of two key components: a retrieval system and a generator. The retrieval system scans through a database of documents to fetch the most relevant ones in response to a user query. The most recent solutions for retrieval systems employed in RAGs rely on semantic search utilizing embeddings.</p><p>The generator, on the other hand, these retrieved documents to generate a well-informed answer. This process ensures that the system provides responses that are both informative and contextually accurate. In the current project setup (April 2024), the gpt-4 model powers the generation of responses. For the creation of embeddings, we utilize text-embedding-ada-002 (https://platform.openai.com/docs/models/embedd ings).</p><p>While our dataset contains about 37000 legal acts, the need for partitioning these laws for a granular retrieval process amplifies the total count of retrievable documents into about 371000 texts ("chunks"). This extensive partitioning provides a more detailed context for the RAG system, allowing for more accurate answer generation. On the other side the increased number of documents naturally presents a challenge for our retrieval process.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Discussion and conclusions</head><p>In this project, we are trying different combinations of the mentioned parameters using both open-source and closed-source models to investigate the readiness of LLMs to build a system for legislative research.</p><p>We are performing two evaluation steps to compare different models and parameters:</p><p>1. Search engine evaluation: we are comparing different Embedding models, chunking strategies and k-nearest neighbors search techniques to select the best combinations to retrieve good-quality results.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Response generator evaluation: having fixed</head><p>the best combination for contextual information retrieval thanks to step 1 evaluation, we will compare the quality of the generated response using different prompt templates, LLMs and LLMs parameters. For step 1 evaluation, we are creating a gold dataset using expert annotators and use standard search engine evaluation metrics such as the Mean Reciprocal Rank. For step 2, evaluation of different settings will be proposed to experts who will ask the same questions and attribute scores to each response. Preliminary results have shown that when the context provided to gpt-4 by the retrieval system is consistent with the question asked, the generated answer is concise, comprehensible, and accurate. This consistency significantly minimizes the problem of hallucination, wherein the model might generate false or nonsensical information.</p><p>The large number of chunks that can contribute to the generation of the answer naturally presents a challenge for our retrieval process. In this context, we are actively exploring strategies to improve the efficiency of this crucial component. One of the promising directions we are considering involves leveraging not only the semantic content of the normative sources but also the boundary information, such as metadata.</p><p>The inclusion of metadata in our retrieval process could potentially imbue our system with the ability to hone in on the most relevant documents, thereby optimizing the retrieval process and improving the overall performance of the chat system. On the other hand, the utilization of a specific embedding model built on legal data could be beneficial, as opposed to a generic embedder. Indeed, this model could provide a more nuanced understanding of the legal texts, thus enhancing the retrieval process.</p><p>The evaluation of the preliminary results is promising: on simple tasks, i.e. simple queries, the system perform on par with human experts, as attested in similar researches <ref type="bibr" target="#b25">[26]</ref>.</p><p>We must highlight the need for legal experts to make qualitative assessments, as well as quantitative evaluations, both due to the complexity of the legal domain and because the evaluation can be done in different ways depending on the use case, aim and user target. It does not seem possible to create a unique dataset to evaluate the quality of the results in an automatic way for QA tasks in the legal domain. Multiple points of view may be equally valid, and a unique 'ground truth' may not exist, as discussed for other tasks in Perspectivist Approaches to NLP <ref type="bibr" target="#b26">[27]</ref> A general problem for experiments with LLMs is "non-repeatability": the experiments are not exactly reproducible because the systems are not deterministic; moreover, in proprietary commercial systems, we do not know the LLM's parameters and the dataset used for training; new versions and new LLMs are released quickly.</p><p>Here are some additional critical issues that we are addressing:</p><p>• text length limitations: currently, there's a limit on the length of text the LLM can handle effectively. This necessitates breaking down longer texts, which can be cumbersome;</p><p>• impact of short queries: the chatbot's accuracy and precision suffer when responding to very short or poorly defined queries. More detailed user queries lead to better results;</p><p>• single vs. multiple documents: the chatbot performs best when responding to queries that target information from a single document, rather than synthesizing information from multiple sources.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head></head><label></label><figDesc>Ital-IA 2024: 4th National Conference on Artificial Intelligence, organized by CINI, May 29-30, 2024, Naples, Italy* Corresponding author.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head></head><label></label><figDesc>questionnaire contained 22 questions: 4 questions for demographic information (age, gender, education, profession); 9 multiple choice questions; 6 open-ended questions; 2 yes/no questions, and 1 rating question. Regarding the topic of the use of LLMs for accessing European laws, the most important questions are: "7) To search for legal documents, regulations and rulings, do you mainly use the EUR-Lex search engine, or do you use Google Search or something else?". "17) In the legal domain, could a generative AI chatbot help search and interaction?". "18) What kind of requests would you make to the chatbot? Write one or more example requests.". "19) Do you know what generative AI is and/or are you a user of generative AI tools?". "21) Do you have any concerns about the use of generative AI in the legal field?".The following table(Table 1) presents the numbers of Submissions, number of people that did not complete the questionnaire (Starts), number of people that viewed the questionnaire (Views) in Italian and English, as of March 30, 2024.</figDesc><table><row><cell>Table 1</cell><cell></cell><cell></cell><cell></cell></row><row><cell cols="4">Questionnaire results in Italy (language: Italian),</cell></row><row><cell cols="4">other EU countries (language: English), and total</cell></row><row><cell>results</cell><cell></cell><cell></cell><cell></cell></row><row><cell>LANG</cell><cell>VIEWS</cell><cell>STARTS</cell><cell>SUBMISSIONS</cell></row><row><cell>Italian</cell><cell>769</cell><cell>315</cell><cell>192</cell></row><row><cell>English</cell><cell>530</cell><cell>184</cell><cell>105</cell></row><row><cell>Total</cell><cell>1299</cell><cell>499</cell><cell>297</cell></row></table></figure>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgements</head><p>Chat-EUR-Lex project is funded within the framework of the NGI Search project under grant agreement No 101069364. Views and opinions expressed are however those of the author(s) only and do not necessarily reflect those of the European Union or European Commission. Neither the European Union nor the granting authority can be held responsible for them.</p></div>
			</div>

			<div type="annex">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>A. Online Resources</head><p>The GitHub project repository can be consulted at https://github.com/Aptus-AI/chat-eur-lex.</p><p>The dataset has been published on https://huggingface.co/datasets/AptusAI/chateur-lex .</p></div>			</div>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<author>
			<persName><forename type="first">J</forename><surname>Cui</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Yan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Yuan</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2306.16092</idno>
		<title level="m">Chatlaw: Open-source legal large language model with integrated external knowledge bases</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b1">
	<monogr>
		<author>
			<persName><forename type="first">J</forename><surname>Lai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Gan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Qi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">S</forename><surname>Yu</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2312.03718</idno>
		<title level="m">Large Language Models in Law: A Survey</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b2">
	<monogr>
		<author>
			<persName><forename type="first">S</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Irsoy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Lu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Dabravolski</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Dredze</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Gehrmann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Kambadur</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Rosenberg</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Mann</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2303.17564v2</idno>
		<title level="m">BloombergGPT: A Large Language Model for Finance</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Evaluating large language models for use in healthcare: A framework for translational value assessment</title>
		<author>
			<persName><forename type="first">S</forename><surname>Reddy</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.imu.2023.101304</idno>
		<ptr target="https://doi.org/10.1016/j.imu.2023.101304" />
	</analytic>
	<monogr>
		<title level="j">Informatics in Medicine Unlocked</title>
		<imprint>
			<biblScope unit="volume">41</biblScope>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Intelligenza artificiale e Diritto. Dai sistemi esperti &quot;classici&quot; ai sistemi esperti &quot;evoluti&quot;: tecnologia e implementazione giuridica</title>
		<author>
			<persName><forename type="first">A</forename><surname>Contaldo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Campara</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Intelligenza artificiale. Algoritmi giuridici. Ius condendum o &quot;fantadiritto</title>
				<editor>
			<persName><forename type="first">G</forename><forename type="middle">Taddei</forename><surname>Elmi</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Contaldo</surname></persName>
		</editor>
		<meeting><address><addrLine>Pisa</addrLine></address></meeting>
		<imprint>
			<publisher>Pacini editore</publisher>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page">24</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<monogr>
		<author>
			<persName><forename type="first">Z</forename><surname>Sun</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2303.09136</idno>
		<title level="m">A short survey of viewing large language models in legal aspect</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b6">
	<monogr>
		<author>
			<persName><forename type="first">G</forename><surname>Finocchiaro</surname></persName>
		</author>
		<title level="m">Artificial intelligence. What are the rules? Il Mulino</title>
				<meeting><address><addrLine>Bologna</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Context-Aware Classification of Legal Document Pages</title>
		<author>
			<persName><forename type="first">P</forename><surname>Fragkogiannis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Forster</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">E</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Zhang</surname></persName>
		</author>
		<idno type="DOI">10.1145/3539618.3591839</idno>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 46th International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR &apos;23)</title>
				<meeting>the 46th International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR &apos;23)<address><addrLine>NY</addrLine></address></meeting>
		<imprint>
			<publisher>Association for Computing Machinery</publisher>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="3285" to="3289" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<monogr>
		<author>
			<persName><forename type="first">D</forename><surname>Trautmann</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2308.04138</idno>
		<title level="m">Large Language Model Prompt Chaining for Long Legal Document Classification</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">The unreasonable effectiveness of large language models in zeroshot semantic annotation of legal texts</title>
		<author>
			<persName><forename type="first">J</forename><surname>Savelka</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><forename type="middle">D</forename><surname>Ashley</surname></persName>
		</author>
		<idno type="DOI">10.3389/frai.2023.1279794</idno>
	</analytic>
	<monogr>
		<title level="j">Frontiers Artificial Intelligence</title>
		<imprint>
			<biblScope unit="volume">6</biblScope>
			<biblScope unit="page" from="1" to="14" />
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">The summarization of legal texts: an experiment with GPT-3</title>
		<author>
			<persName><forename type="first">M</forename><surname>Cherubini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Romano</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Bolioli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>De Francesco</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Benedetto</surname></persName>
		</author>
		<idno type="DOI">10.32091/RIID0103</idno>
		<ptr target="https://doi.org/10.32091/RIID0103" />
	</analytic>
	<monogr>
		<title level="j">Rivista italiana di informatica e diritto</title>
		<imprint>
			<biblScope unit="volume">5</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="191" to="204" />
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<monogr>
		<title level="m" type="main">MILDSum: A Novel Benchmark Dataset for Multilingual Summarization of Indian Legal Case Judgments</title>
		<author>
			<persName><forename type="first">D</forename><surname>Datta</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Soni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Mukherjee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Ghosh</surname></persName>
		</author>
		<idno type="DOI">10.48550/arXiv.2310.18600</idno>
		<idno type="arXiv">arXiv:2310.18600v1</idno>
		<idno>.18600</idno>
		<ptr target="https://doi.org/10.48550/arXiv.2310" />
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Fine-tuning GPT-3 for legal rule classification</title>
		<author>
			<persName><forename type="first">D</forename><surname>Liga</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Robaldo</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.clsr.2023.105864</idno>
	</analytic>
	<monogr>
		<title level="j">Computer Law &amp; Security Review</title>
		<imprint>
			<biblScope unit="volume">51</biblScope>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Pre-trained language models for the legal domain: a case study on Indian law</title>
		<author>
			<persName><forename type="first">S</forename><surname>Paul</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Mandal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Goyal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Ghosh</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Nineteenth International Conference on Artificial Intelligence and Law</title>
				<meeting>the Nineteenth International Conference on Artificial Intelligence and Law</meeting>
		<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="187" to="196" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Performance analysis of large language models in the domain of legal argument mining</title>
		<author>
			<persName><forename type="first">A</forename><surname>Al Zubaer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Granitzer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Mitrović</surname></persName>
		</author>
		<idno type="DOI">10.3389/frai.2023.1278796</idno>
	</analytic>
	<monogr>
		<title level="j">Frontiers Artificial Intelligence</title>
		<imprint>
			<biblScope unit="volume">6</biblScope>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<monogr>
		<author>
			<persName><forename type="first">J</forename><surname>Savelka</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><forename type="middle">D</forename><surname>Ashley</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">A</forename><surname>Gray</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Westermann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Xu</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2306.09525v2</idno>
		<title level="m">Explaining Legal Concepts with Augmented Large Language Models (GPT-4)</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">ai: Legal Generative AI for Regulatory Compliance</title>
		<author>
			<persName><forename type="first">J</forename><surname>Ioannidis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Harper</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">S</forename><surname>Quah</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Hunter</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Gracenote</forename></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Third International Workshop on Artificial Intelligence and Intelligent Assistance for Legal Professionals in the Digital Workplace (LegalAIIA 2023) co-located with (ICAIL 2023)</title>
		<title level="s">CEUR-WS</title>
		<meeting>the Third International Workshop on Artificial Intelligence and Intelligent Assistance for Legal Professionals in the Digital Workplace (LegalAIIA 2023) co-located with (ICAIL 2023)</meeting>
		<imprint>
			<publisher>Elsevier</publisher>
			<biblScope unit="page" from="20" to="31" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<monogr>
		<title level="m" type="main">Large Language Models and the Future of Law</title>
		<author>
			<persName><forename type="first">D</forename><surname>Charlotin</surname></persName>
		</author>
		<ptr target="https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4548258" />
		<imprint>
			<date type="published" when="2023-04-04">2023. Accessed 04 april, 2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">Can GPT-3 Perform Statutory Reasoning?</title>
		<author>
			<persName><forename type="first">A</forename><surname>Blair-Stanek</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Holzenberger</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Van Durme</surname></persName>
		</author>
		<idno type="DOI">10.1145/3594536.3595163</idno>
	</analytic>
	<monogr>
		<title level="m">The Nineteenth International Conference on Artificial Intelligence and Law (ICAIL 2023)</title>
				<meeting><address><addrLine>NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="22" to="31" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">The emerging artificial intelligence (AI) and national uniform legislation</title>
		<author>
			<persName><forename type="first">G</forename><surname>Hill</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Australian Law Journal</title>
		<imprint>
			<biblScope unit="volume">97</biblScope>
			<biblScope unit="issue">5</biblScope>
			<biblScope unit="page" from="303" to="306" />
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<monogr>
		<author>
			<persName><forename type="first">L</forename><surname>Huang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Yu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Ma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Zhong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Feng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Liu</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2311.05232</idno>
		<title level="m">A survey on hallucination in large language models: Principles, taxonomy, challenges, and open questions</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b21">
	<monogr>
		<title level="m" type="main">A comprehensive survey of hallucination mitigation techniques in large language models</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">M</forename><surname>Tonmoy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">M</forename><surname>Zaman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Jain</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Rani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Rawte</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Chadha</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Das</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2401.01313</idno>
		<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">Retrievalaugmented generation for knowledge-intensive nlp tasks</title>
		<author>
			<persName><forename type="first">P</forename><surname>Lewis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Perez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Piktus</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Petroni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Karpukhin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Goya</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Kiela</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">of Neural Information Processing Systems</title>
				<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="volume">33</biblScope>
			<biblScope unit="page" from="9459" to="9474" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<monogr>
		<title level="m" type="main">Parameter-efficient fine-tuning methods for pretrained language models: A critical review and assessment</title>
		<author>
			<persName><forename type="first">L</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Xie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">Z J</forename><surname>Qin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Tao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">L</forename><surname>Wang</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2312.12148</idno>
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b24">
	<monogr>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">J</forename><surname>Hu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Shen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Wallis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Allen-Zhu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Chen</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2106.09685</idno>
		<title level="m">Lora: Low-rank adaptation of large language models</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b25">
	<analytic>
		<title level="a" type="main">Benchmarking large language models for news summarization</title>
		<author>
			<persName><forename type="first">T</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Ladhak</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Durmus</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Liang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Mckeown</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">B</forename><surname>Hashimoto</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Transactions of the Association for Computational Linguistics</title>
		<imprint>
			<biblScope unit="volume">12</biblScope>
			<biblScope unit="page" from="39" to="57" />
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b26">
	<analytic>
		<title level="a" type="main">Toward a perspectivist turn in ground truthing for predictive computing</title>
		<author>
			<persName><forename type="first">F</forename><surname>Cabitza</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Campagner</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Basile</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the AAAI Conference on Artificial Intelligence</title>
				<meeting>the AAAI Conference on Artificial Intelligence</meeting>
		<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="volume">37</biblScope>
			<biblScope unit="page" from="6860" to="6868" />
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
