<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">SAVIA: Artificial Intelligence in support of the lawmaking process</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Michele</forename><surname>Visciarelli</surname></persName>
							<email>m.visciarelli@cineca.it</email>
							<affiliation key="aff0">
								<orgName type="institution">CINECA</orgName>
								<address>
									<addrLine>via Magnanelli 6/3, Casalecchio di Reno (B0)</addrLine>
									<postCode>40033</postCode>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Giovanni</forename><surname>Guidi</surname></persName>
							<email>g.guidi@cineca.it</email>
							<affiliation key="aff0">
								<orgName type="institution">CINECA</orgName>
								<address>
									<addrLine>via Magnanelli 6/3, Casalecchio di Reno (B0)</addrLine>
									<postCode>40033</postCode>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Laura</forename><surname>Morselli</surname></persName>
							<email>l.morselli@cineca.it</email>
							<affiliation key="aff0">
								<orgName type="institution">CINECA</orgName>
								<address>
									<addrLine>via Magnanelli 6/3, Casalecchio di Reno (B0)</addrLine>
									<postCode>40033</postCode>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Domitilla</forename><surname>Brandoni</surname></persName>
							<email>d.brandoni@cineca.it</email>
							<affiliation key="aff0">
								<orgName type="institution">CINECA</orgName>
								<address>
									<addrLine>via Magnanelli 6/3, Casalecchio di Reno (B0)</addrLine>
									<postCode>40033</postCode>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Giuseppe</forename><surname>Fiameni</surname></persName>
							<email>gfiameni@nvidia.com</email>
							<affiliation key="aff1">
								<orgName type="institution">NVIDIA AI Technology Center</orgName>
								<address>
									<settlement>Milan</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Luisa</forename><surname>Monti</surname></persName>
							<email>luisa.monti@regione.emilia-romagna.it</email>
							<affiliation key="aff2">
								<orgName type="institution">Assemblea Legislativa Emilia Romagna</orgName>
								<address>
									<addrLine>viale Aldo Moro 50</addrLine>
									<postCode>41127</postCode>
									<settlement>Bologna</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Stefano</forename><surname>Bianchini</surname></persName>
							<email>stefano.bianchini@regione.emilia-romagna.it</email>
							<affiliation key="aff2">
								<orgName type="institution">Assemblea Legislativa Emilia Romagna</orgName>
								<address>
									<addrLine>viale Aldo Moro 50</addrLine>
									<postCode>41127</postCode>
									<settlement>Bologna</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Cosimo</forename><surname>Tommasi</surname></persName>
							<email>cosimo.tommasi@regione.emilia-romagna.it</email>
							<affiliation key="aff2">
								<orgName type="institution">Assemblea Legislativa Emilia Romagna</orgName>
								<address>
									<addrLine>viale Aldo Moro 50</addrLine>
									<postCode>41127</postCode>
									<settlement>Bologna</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">SAVIA: Artificial Intelligence in support of the lawmaking process</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">F33854F0B35DABEE4561F46A7B498863</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T16:57+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Generative AI, LLM, Legal AI, NLP (C. Tommasi) 0000-0003-0753-2571 (M. Visciarelli)</term>
					<term>0000-XXX (G. Guidi)</term>
					<term>0000-0003-0753-2571 (L. Morselli)</term>
					<term>0000-0002-8157-1459 (D. Brandoni)</term>
					<term>0000-0001-8687-6609 (G. Fiameni)</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>We explore the use of open-source Large Language Models (LLMs) to support legal professionals, lawmakers, and citizens in accessing information on the current and past legislation of the Emilia-Romagna region. We develop a generative AI tool based on the Retrieval-Augmented Generation (RAG) technique to answer questions related to regional laws and their implementing acts, retrieving relevant information from the Emilia-Romagna law corpus. To adapt pre-trained LLMs to this downstream task, we follow a multi-step approach. First, we use the QLoRa technique to quantize and adapt the pre-trained LLMs to the regional legal text dataset. Next, we fine-tune the domain-adapted models using an "ad-hoc" instruction-based dataset. We then implement a module to retrieve relevant contextual information from the legal documents dataset. Finally, we align the models with domain-specific instructions using RAG-based prompting. We evaluate the performance of the domain-adapted models using the perplexity metric, and the results of the final fine-tuned models are assessed by domain experts, focusing on the quality of the generated text and the relevance of the answers. Our results show that domain adaptation on domain-specific text is a crucial step for enhancing the quality of the generated text in expert domains, such as legal texts, which contain a vast amount of specialized vocabulary and expressions. This approach leads to higher performance compared to models fine-tuned only on small Question-Answer datasets. Additionally, our findings highlight the importance of the retrieval module, which must be able to reliably find the most relevant documents to provide useful and up-to-date insights to lawmakers and citizens.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>In the last years the interest for Generative Artificial Intelligence (Generative AI) applications has grown importance among the research and industry community, thanks to the introduction of Foundation Models in different AI domains, such as text generation (GPT-series <ref type="bibr" target="#b0">[1,</ref><ref type="bibr" target="#b1">2,</ref><ref type="bibr" target="#b2">3]</ref>, LLaMA-series <ref type="bibr" target="#b3">[4,</ref><ref type="bibr" target="#b4">5,</ref><ref type="bibr" target="#b5">6]</ref>, MEGATRON <ref type="bibr" target="#b6">[7]</ref>), image generation (DALL• E <ref type="bibr" target="#b7">[8,</ref><ref type="bibr" target="#b8">9,</ref><ref type="bibr" target="#b9">10]</ref>, Stable Diffusion <ref type="bibr" target="#b10">[11]</ref>), and video generation (Sora <ref type="bibr" target="#b11">[12]</ref>). The progress in Deep Learning modelling has been fostered by important advancements in the neural network (NN) research, such as the introduction of the Transformer architecture in Nat-ural Language Processing (NLP) <ref type="bibr" target="#b12">[13]</ref>, by improvements in hardware acceleration for linear algebra, expanding model's size up to several billions of parameters, by the introduction of quantization techniques allowing training of large NNs even on consumer GPUs <ref type="bibr" target="#b13">[14,</ref><ref type="bibr" target="#b14">15]</ref>, and by the release of large high-quality and open-source datasets <ref type="bibr" target="#b15">[16]</ref>.</p><p>Large Language Models (LLMs) for text generation have achieved remarkable performance and great interest even outside the research and industry professional community, in particular after the release of ChatGPT to the public <ref type="bibr" target="#b16">[17]</ref>. Despite their success, the use of LLMs on domain-specific Question-Answer (QA) tasks still face several challenges, that hinder their spread beyond the research community, especially when applied to tasks for which explainability and high quality responses are of paramount importance <ref type="bibr" target="#b17">[18]</ref>. Some of the challenges that LLMs are still facing are the following:</p><p>• difficulty of maintaining up-to-date knowledge • costs of training and inference of large models, costs and difficulty to collect large amount of highquality domain-specific data • hallucinated answers, i.e. answers that provided false information without warning</p><p>• out-of-date or generic answers, even when the user expects a specific, current response Retrieval-Augmented Generation or RAG has recently emerged as a paradigm to address such challenges <ref type="bibr" target="#b18">[19]</ref>. In particular, RAG combines a language model with an information retrieval system to dynamically fetch relevant external information to enhance the model's responses, by encoding the user's question into a dense representation, and retrieving passages relevant to the question from an indexed data source, adding this information to the LLM prompt. Different studies have shown that RAG enhances the quality of the generation process, leading to higher accuracy, better robustness, reduced hallucinations, higher interpretability, and even the possibility to perform open-domain QA just by updating the knowledge-base <ref type="bibr" target="#b19">[20,</ref><ref type="bibr" target="#b20">21]</ref>. RAG also offers a balanced approach in terms of customization and resource requirements, being more flexible and cost-effective than full fine-tuning, although still requiring labeled data and a supervised training phase.</p><p>In this work we present SAVIA, a project developed by CINECA and Assemblea Legislativa of Emilia-Romagna. This project, started in Autumn 2023 and expected to end in march 2025, has the goal of creating a model capable of answering questions on the Region's law and their respective implementing acts, as well as on related "exante" and "ex-post" reports of the laws' impact. In Section 2 we present the data used for this project and the workflow that has been adopted. In Section 3, we describe the procedure and the details of the experiments and tests conducted, and in Section 4 we show the obtained results. Our conclusion are then presented in Section 5.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Methodology</head><p>To obtain a model capable of understanding Italian language in the law domain and responding to questions related to laws enacted in the Emilia-Romagna region, we followed a multi-step approach. We started from an open-source LLM and adapted it to the legal language through unsupervised domain adaptation (Section 2.1). The resulting domain-adapted model was then fine-tuned for question-answering (Q&amp;A) on an instruction-based dataset prepared by domain experts for this purpose (Section 2.2). Finally, we implemented a domain-adapted retrieval model (Section 2.3) to enrich the answers with relevant information from the law corpus.</p><p>The full workflow was reproduced starting from different open-source LLMs:</p><p>• LLaMAntino-2-7b-hf-ITA: a 7B model, based on LLama-2, specifically fine-tuned for the Italian language <ref type="bibr" target="#b21">[22]</ref>.</p><p>• Mistral-7B-v0.1: a 7B model, that implements grouped-query and sliding window attention, Rotary Position Embedding, that can handle context of arbitrary size <ref type="bibr" target="#b22">[23]</ref>. • Mixtral-8x7B-Instruct-v0.1: a 46.7B mixture of experts model, trained on instructions in English, French, Italian, German and Spanish, with maximum context length of 32k <ref type="bibr" target="#b23">[24]</ref>.</p><p>Domain experts qualitatively evaluated the performances of the final models obtained from the different pre-trained LLMs.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1.">Unsupervised Domain-Adaptation</head><p>The first step in the procedure was the domain-adaptation of the model on legal text. We collected the PDFs of the regional laws of Emilia-Romagna, as well as the relative implementing acts at the regional level (e.g. "atto del dirigente" and "atto di giunta") and the available reports on the expected and measured impact of a given law (e.g. "clausola valutativa", "ex-ante and "ex-post" reports). We split the legal documents in chunks, and we implemented a cleaning pipeline to remove typos, bad characters, and irrelevant parts of the documents such as headers and footers. We also added mii-llm/gazzetta-ufficiale <ref type="bibr" target="#b24">[25]</ref> in the training dataset, given the affinity of this dataset to our application, both in language, semantics and type of documents. We did not perform domain-adaptive tokenization <ref type="bibr" target="#b25">[26]</ref>, using instead the pre-trained models native tokenizers to tokenize the legal corpus.</p><p>Not all the three model under investigation underwent domain adaptation. LLaMAntino-2-7b-hf-ITA and Mistral-7B-v0.1 were adapted, while Mixtral-8x7B-Instruct-v0.1, after tests regarding its native capabilities of producing adequate Italian legal text, has not been domain adapted.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.">Model Alignment on Instruction-Based dataset</head><p>With the support of domain experts, we generated an Q&amp;A dataset mimicking different levels of domain language proficiency, ranging from questions that could by written by non-expert users, to the ones that may be asked by experts in the legal domain. We developed a semi-automatic procedure to further enrich this Q&amp;A dataset, using legal documents metadata. The following is an example included in the instruction-based dataset:</p><p>• Q: "Da quando è stata istituita la regione, quali normative sono state adottate per incentivare la partecipazione?" • A: "La prima legge regionale riguardante la partecipazione ad essere stata approvata è la legge numero 3 del 2010. In seguito, la legge numero 3 del 2010 è stata abolita e sostituita con la legge regionale numero 15 del 2018. ", To fine-tune the domain-adapted LLMs, we used the instruction-based dataset prepared by the domain experts. For the loss function computation, we removed the portion of the text containing the prompt, as in many cases the prompt added by the RAG module can account for up to 50% of the total text length. This approach helped to optimize the training process more effectively.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.3.">Domain-Adapted Retrieval Model</head><p>To enrich the user's question with relevant information from the legal documents database, we develop a retrieval module based on Semantic Similarity search technique. We used a Sentence-BERT model <ref type="bibr" target="#b26">[27]</ref> to populate a vector store with embedding generated from the legal documents text chunks. The content similar to a user's question is retrieved using the semantic search library FAISS <ref type="bibr" target="#b27">[28,</ref><ref type="bibr" target="#b28">29]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Experiment</head><p>The project has been carried out exploiting the computational resources of the supercomputer LEONARDO, hosted by CINECA. Each node in the booster partition is equipped with four NVidia A100 SXM6 64GB GPUs and a single 32-cores Intel Ice Lake CPU.</p><p>For all models, only data parallelism has been employed, given the fact that all these models could adequately fit in the VRAM of the GPUs at our disposal. For the same reason, LLaMAntino-2-7b-hf-ITA and Mistral-7B-v0.1 have not been quantized during domain adaptation and instruction fine-tuning, opting to preserve the weights' precision. Mixtral-8x7B-Instruct-v0.1 underwent 4-bit quantization instead <ref type="bibr" target="#b29">[30]</ref>, due to its size. For domain adaptation and instruction fine-tuning, we applied LoRA adapters on Q, K, V layers of the models <ref type="bibr" target="#b14">[15]</ref>. The training procedure for the models under study were the following:</p><p>• pre-trained LLMs causal Language Modelling on the legal text chunks. This has been performed on LLaMAntino-2-7b-hf-ITA and Mistral-7B-v0. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Results</head><p>To evaluate the quality and proceed to select the candidate for instruction-based fine-tuning, all domainadapted models were evaluated using the perplexity metric (PPL) on an held-out evaluation dataset based on laws.</p><p>The metric is reported in Table <ref type="table" target="#tab_0">1</ref>.</p><p>Three different domain experts (lawmakers of Assemblea Legislativa) were asked to evaluate the answers generated by the final instruction fine-tuned models to a set of 25 questions. The qualitative analysis of the experts reported that, in general, the answers provided by the LLaMAntino-based model were considered too short and dry. The answers provided by the Mixtral-based model were considered the most complete, clear and satisfactory in terms of quality of the used specific words. Below we report an example of answers provided by the different final models on a given question. For context, we also include the answer of chatGPT (3.5) to the same question.</p><p>• Question: Sul tema della partecipazione, quali leggi sono state fatte in Emilia-Romagna? • Answer of Mixtral-8x7B-Instruct-v0.1 fine-tuned:</p><p>La prima legge regionale approvata in tema di partecipazione è la legge regionale 9 febbraio 2010, n. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Conclusions</head><p>We explored different approaches to adapt open-source LLMs for question-answering on the Emilia-Romagna law corpus. We adapted the different LLMs on a corpus composed of the Emilia-Romagna regional laws and the relative implementing acts, and we further refined the domain-adapted models on a custom QA dataset provided by domain experts. Finally, we exploited RAG to enrich the user's question with relevant contextual information extracted from the law database. We experimented with different open-source LLMs, such as Mistral-7B-v0.1, LLaMAntino-2-7b-hf-ITA, Mixtral-8x7B-Instruct-v0.1. Our results show that domain-adapted LLMs that are able to answer specific domain questions can be a helpful tool to support decisionmaking in specialized fields such as the legal domain, that often need to retrieve exact, concise and easy-tounderstand information from large and unstructured data sources.</p><p>Given the scope and the length of the project, several improvements to the workflow are foreseen in the near future, as well as the possibility to test with more pre-trained open source models, for example new Italiannative models that will be developed in the near future and domain adaptation of Mixture-of-Experts models (such as Mixtral-8x7B-v0.1). Our future work will also focus on further improving the retrieval module with better embedding models, and on applying more powerful techniques to train the LLMs, such as Direct Preference Optimization (DPO, <ref type="bibr" target="#b30">[31]</ref>) and Reinforcement Learning from Human Feedback (RLHF, <ref type="bibr" target="#b31">[32]</ref>).</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1</head><label>1</label><figDesc>Perplexity for base and domain-adapted models under study.</figDesc><table><row><cell>1 and needed for each model, on average,</cell></row><row><cell>400 GPU hours (approximately 4 days on a sin-</cell></row><row><cell>gle LEONARDO booster node) training for four</cell></row><row><cell>epochs;</cell></row><row><cell>• model alignment on domain adapted</cell></row><row><cell>LLaMAntino-2-7b-hf-ITA and Mistral-7B-v0.1,</cell></row><row><cell>and on base pre-trained Mixtral-8x7B-Instruct-</cell></row><row><cell>v0.1, on the QA dataset. This step required</cell></row><row><cell>approximately 96 GPU hours, or 24 node hours</cell></row><row><cell>(4 GPUs per node), to complete a 12 epochs</cell></row><row><cell>training run, on a single LEONARDO node.</cell></row></table></figure>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgments</head><p>We are extremely grateful to the President of Assemblea Legislativa Emilia-Romagna, Emma Petitti, for the long-eyed vision that created the conditions to launch the project -and to the Director General of Assemblea Legislativa Emilia-Romagna, Leonardo Draghetti, to set strategically the project and ensure the necessary human and material resources.</p><p>Besides, this endeavour would not have been possible without the commitment of the President of CINECA, Francesco Ubertini, and of the Director of the Supercomupting applications and innovation Director of CINECA, Sanzio Bassini.</p><p>Special thanks goes to Giovanna Favero of Assemblea Legislativa Emilia-Romagna for her efforts in making available laws, implementing acts, as well as related "exante" and "ex-post" reports.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">B</forename><surname>Brown</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Mann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Ryder</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Subbiah</surname></persName>
		</author>
		<idno type="DOI">10.48550/arXiv.2005.14165</idno>
		<idno type="arXiv">arXiv:2005.14165arXiv:2005.14165</idno>
		<title level="m">Language Models are Few-Shot Learners</title>
				<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<monogr>
		<author>
			<persName><forename type="first">M</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Tworek</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Jun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Yuan</surname></persName>
		</author>
		<idno type="DOI">10.48550/arXiv.2107.03374</idno>
		<idno type="arXiv">arXiv:2107.03374arXiv:2107.03374</idno>
		<title level="m">Evaluating Large Language Models Trained on Code</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<monogr>
		<title/>
		<author>
			<persName><forename type="first">J</forename><surname>Openai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Achiam</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Adler</surname></persName>
		</author>
		<author>
			<persName><surname>Agarwal</surname></persName>
		</author>
		<idno type="DOI">10.48550/arXiv.2303.08774</idno>
		<idno type="arXiv">arXiv:2303.08774arXiv:2303.08774</idno>
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
	<note type="report_type">GPT-4 Technical Report</note>
</biblStruct>

<biblStruct xml:id="b3">
	<monogr>
		<author>
			<persName><forename type="first">H</forename><surname>Touvron</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Lavril</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Izacard</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Martinet</surname></persName>
		</author>
		<idno type="DOI">10.48550/arXiv.2302.13971</idno>
		<idno type="arXiv">arXiv:2302.13971arXiv:2302.13971</idno>
		<title level="m">LLaMA: Open and Efficient Foundation Language Models</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<monogr>
		<author>
			<persName><forename type="first">H</forename><surname>Touvron</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Martin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Stone</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Albert</surname></persName>
		</author>
		<idno type="DOI">10.48550/arXiv.2307.09288</idno>
		<idno type="arXiv">arXiv:2307.09288arXiv:2307.09288</idno>
		<title level="m">Llama 2: Open Foundation and Fine-Tuned Chat Models</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<monogr>
		<author>
			<persName><forename type="first">B</forename><surname>Rozière</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Gehring</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Gloeckle</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Sootla</surname></persName>
		</author>
		<idno type="DOI">10.48550/arXiv.2308.12950</idno>
		<idno type="arXiv">arXiv:2308.12950arXiv:2308.12950</idno>
		<title level="m">Code Llama: Open Foundation Models for Code</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<monogr>
		<author>
			<persName><forename type="first">D</forename><surname>Narayanan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Shoeybi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Casper</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Legresley</surname></persName>
		</author>
		<idno type="DOI">10.48550/arXiv.2104.04473</idno>
		<idno type="arXiv">arXiv:2104.04473arXiv:2104.04473</idno>
		<title level="m">Efficient Large-Scale Language Model Training on GPU Clusters Using Megatron-LM</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv e-prints</note>
</biblStruct>

<biblStruct xml:id="b7">
	<monogr>
		<author>
			<persName><forename type="first">A</forename><surname>Ramesh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Pavlov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Goh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Gray</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Voss</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Radford</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Sutskever</surname></persName>
		</author>
		<idno type="DOI">10.48550/arXiv.2102.12092</idno>
		<idno type="arXiv">arXiv:2102.12092arXiv:2102.12092</idno>
		<title level="m">Zero-Shot Text-to-Image Generation</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<monogr>
		<author>
			<persName><forename type="first">A</forename><surname>Ramesh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Dhariwal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Nichol</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Chu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Chen</surname></persName>
		</author>
		<idno type="DOI">10.48550/arXiv.2204.06125</idno>
		<idno type="arXiv">arXiv:2204.06125arXiv:2204.06125</idno>
		<title level="m">Hierarchical Text-Conditional Image Generation with CLIP Latents</title>
				<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv e-prints</note>
</biblStruct>

<biblStruct xml:id="b9">
	<monogr>
		<author>
			<persName><forename type="first">Z</forename><surname>Shi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Zhou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Qiu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Zhu</surname></persName>
		</author>
		<idno type="DOI">10.48550/arXiv.2006.11807</idno>
		<idno type="arXiv">arXiv:2006.11807arXiv:2006.11807</idno>
		<title level="m">Improving Image Captioning with Better Use of Captions</title>
				<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<monogr>
		<author>
			<persName><forename type="first">R</forename><surname>Rombach</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Blattmann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Lorenz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Esser</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Ommer</surname></persName>
		</author>
		<idno type="DOI">10.48550/arXiv.2112.10752</idno>
		<idno type="arXiv">arXiv:2112.10752arXiv:2112.10752</idno>
		<title level="m">High-Resolution Image Synthesis with Latent Diffusion Models</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<monogr>
		<title level="m" type="main">Video generation models as world simulators</title>
		<author>
			<persName><surname>Openai</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2024">2024</date>
			<pubPlace>OpenAI</pubPlace>
		</imprint>
	</monogr>
	<note type="report_type">Tech. rep.</note>
</biblStruct>

<biblStruct xml:id="b12">
	<monogr>
		<author>
			<persName><forename type="first">A</forename><surname>Vaswani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Shazeer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Parmar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Uszkoreit</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Jones</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">N</forename><surname>Gomez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Kaiser</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Polosukhin</surname></persName>
		</author>
		<idno type="DOI">10.48550/arXiv.1706.03762</idno>
		<idno type="arXiv">arXiv:1706.03762arXiv:1706.03762</idno>
		<title level="m">Attention Is All You Need</title>
				<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<monogr>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">J</forename><surname>Hu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Shen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Wallis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Allen-Zhu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Chen</surname></persName>
		</author>
		<idno type="DOI">10.48550/arXiv.2106.09685</idno>
		<idno type="arXiv">arXiv:2106.09685arXiv:2106.09685</idno>
		<title level="m">LoRA: Low-Rank Adaptation of Large Language Models</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<monogr>
		<author>
			<persName><forename type="first">T</forename><surname>Dettmers</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Pagnoni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Holtzman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Zettlemoyer</surname></persName>
		</author>
		<idno type="DOI">10.48550/arXiv.2305.14314</idno>
		<idno type="arXiv">arXiv:2305.14314arXiv:2305.14314</idno>
		<title level="m">QLoRA: Efficient Finetuning of Quantized LLMs</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<monogr>
		<author>
			<persName><forename type="first">H</forename><surname>Face</surname></persName>
		</author>
		<ptr target="https://huggingface.co/datasets" />
		<title level="m">Hugging face datasets</title>
				<imprint>
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<monogr>
		<author>
			<persName><surname>Openai</surname></persName>
		</author>
		<title level="m">Introducing chatgpt</title>
				<meeting><address><addrLine>OpenAI</addrLine></address></meeting>
		<imprint>
			<publisher>Tech. rep</publisher>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<monogr>
		<author>
			<persName><forename type="first">L</forename><surname>Huang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Yu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Ma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Zhong</surname></persName>
		</author>
		<idno type="DOI">10.48550/arXiv.2311.05232</idno>
		<idno type="arXiv">arXiv:2311.05232arXiv:2311.05232</idno>
		<title level="m">A Survey on Hallucination in Large Language Models: Principles, Taxonomy, Challenges, and Open Questions</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<monogr>
		<author>
			<persName><forename type="first">P</forename><surname>Lewis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Perez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Piktus</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Petroni</surname></persName>
		</author>
		<idno type="DOI">10.48550/arXiv.2005.11401</idno>
		<idno type="arXiv">arXiv:2005.11401arXiv:2005.11401</idno>
		<title level="m">Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks</title>
				<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<monogr>
		<author>
			<persName><forename type="first">S</forename><surname>Siriwardhana</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Weerasekera</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Wen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Kaluarachchi</surname></persName>
		</author>
		<idno type="DOI">10.48550/arXiv.2210.02627</idno>
		<idno type="arXiv">arXiv:2210.02627arXiv:2210.02627</idno>
		<title level="m">Improving the Domain Adaptation of Retrieval Augmented Generation (RAG) Models for Open Domain Question Answering</title>
				<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<monogr>
		<author>
			<persName><forename type="first">P</forename><surname>Zhao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Yu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Wang</surname></persName>
		</author>
		<idno type="DOI">10.48550/arXiv.2402.19473</idno>
		<idno type="arXiv">arXiv:2402.19473arXiv:2402.19473</idno>
		<title level="m">Retrieval-Augmented Generation for AI-Generated Content: A Survey</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<monogr>
		<author>
			<persName><forename type="first">P</forename><surname>Basile</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Musacchio</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Polignano</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Siciliani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Fiameni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Semeraro</surname></persName>
		</author>
		<idno type="DOI">10.48550/arXiv.2312.09993</idno>
		<idno type="arXiv">arXiv:2312.09993arXiv:2312.09993</idno>
		<title level="m">LLaMAntino: LLaMA 2 Models for Effective Text Generation in Italian Language</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<monogr>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">Q</forename><surname>Jiang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Sablayrolles</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Mensch</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Bamford</surname></persName>
		</author>
		<idno type="DOI">10.48550/arXiv.2310.06825</idno>
		<idno type="arXiv">arXiv:2310.06825arXiv:2310.06825</idno>
		<title level="m">Mistral 7B</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv e-prints</note>
</biblStruct>

<biblStruct xml:id="b23">
	<monogr>
		<title level="m" type="main">Mixtral of experts</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">A</forename><surname>Team</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2023">2023</date>
			<publisher>Mistral AI</publisher>
		</imprint>
	</monogr>
	<note type="report_type">Tech. rep.</note>
</biblStruct>

<biblStruct xml:id="b24">
	<monogr>
		<author>
			<persName><forename type="first">E</forename><surname>Federici</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Ferraretto</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Landro</surname></persName>
		</author>
		<ptr target="https://huggingface.co/datasets/mii-llm/gazzetta-ufficiale" />
		<title level="m">Gazzetta Ufficiale: A dataset of legislative texts, public and private acts</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b25">
	<monogr>
		<author>
			<persName><forename type="first">M</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T.-D</forename><surname>Ene</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Kirby</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Cheng</surname></persName>
		</author>
		<idno type="DOI">10.48550/arXiv.2311.00176</idno>
		<idno type="arXiv">arXiv:2311.00176arXiv:2311.00176</idno>
		<title level="m">ChipNeMo: Domain-Adapted LLMs for Chip Design</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b26">
	<monogr>
		<author>
			<persName><forename type="first">N</forename><surname>Reimers</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Gurevych</surname></persName>
		</author>
		<idno type="DOI">10.48550/arXiv.1908.10084</idno>
		<idno type="arXiv">arXiv:1908.10084arXiv:1908.10084</idno>
		<title level="m">Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks</title>
				<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b27">
	<analytic>
		<title level="a" type="main">Billion-scale similarity search with GPUs</title>
		<author>
			<persName><forename type="first">J</forename><surname>Johnson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Douze</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Jégou</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Big Data</title>
		<imprint>
			<biblScope unit="volume">7</biblScope>
			<biblScope unit="issue">3</biblScope>
			<biblScope unit="page" from="535" to="547" />
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b28">
	<monogr>
		<author>
			<persName><forename type="first">M</forename><surname>Douze</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Guzhva</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Deng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Johnson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Szilvasy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P.-E</forename><surname>Mazaré</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Lomeli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Hosseini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Jégou</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2401.08281</idno>
		<title level="m">The faiss library</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b29">
	<monogr>
		<title/>
		<author>
			<persName><forename type="first">H</forename><surname>Face</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2023">2023</date>
			<publisher>bitsandbytes, Tech. rep., Hugging Face</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b30">
	<monogr>
		<author>
			<persName><forename type="first">R</forename><surname>Rafailov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Sharma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Mitchell</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Ermon</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">D</forename><surname>Manning</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Finn</surname></persName>
		</author>
		<idno type="DOI">10.48550/arXiv.2305.18290</idno>
		<idno type="arXiv">arXiv:2305.18290arXiv:2305.18290</idno>
		<title level="m">Direct Preference Optimization: Your Language Model is Secretly a Reward Model</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b31">
	<monogr>
		<author>
			<persName><forename type="first">L</forename><surname>Ouyang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Jiang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Almeida</surname></persName>
		</author>
		<idno type="DOI">10.48550/arXiv.2203.02155</idno>
		<idno type="arXiv">arXiv:2203.02155arXiv:2203.02155</idno>
		<title level="m">Training language models to follow instructions with human feedback</title>
				<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
