<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Automatic Summarization of Legal Texts, Extractive Summarization using LLMs</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">David</forename><surname>Preti</surname></persName>
							<email>d.preti@almawave.it</email>
							<affiliation key="aff0">
								<orgName type="institution">Almawave S.P.A</orgName>
								<address>
									<addrLine>via Casal Boccone 10</addrLine>
									<postCode>00133</postCode>
									<settlement>Roma</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Cristina</forename><surname>Giannone</surname></persName>
							<email>c.giannone@almawave.it</email>
							<affiliation key="aff0">
								<orgName type="institution">Almawave S.P.A</orgName>
								<address>
									<addrLine>via Casal Boccone 10</addrLine>
									<postCode>00133</postCode>
									<settlement>Roma</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Andrea</forename><surname>Favalli</surname></persName>
							<email>a.favalli@almawave.it</email>
							<affiliation key="aff0">
								<orgName type="institution">Almawave S.P.A</orgName>
								<address>
									<addrLine>via Casal Boccone 10</addrLine>
									<postCode>00133</postCode>
									<settlement>Roma</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Raniero</forename><surname>Romagnoli</surname></persName>
							<email>r.romagnoli@almawave.it</email>
							<affiliation key="aff0">
								<orgName type="institution">Almawave S.P.A</orgName>
								<address>
									<addrLine>via Casal Boccone 10</addrLine>
									<postCode>00133</postCode>
									<settlement>Roma</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Automatic Summarization of Legal Texts, Extractive Summarization using LLMs</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">0281742352F44890A90454C3179B951B</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T16:57+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Legal Text</term>
					<term>Summarization</term>
					<term>LLM</term>
					<term>Generative AI</term>
					<term>Human in the Loop</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>In this work, we describe the first results of experimentation with summarization systems based on large language models to produce an extractive summarization of the judgments (massime). We propose a novel approach for this task, exploiting the generative capabilities of LLM and removing all possibilities of hallucination. Our study aims to assess the effectiveness and efficiency of generative models in summarizing the court's decisions. Through a comprehensive analysis of several summarization system setups, we evaluate the quality of the summaries generated by each approach and their ability to capture the key legal principles and linguistic features in the courts' decisions.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Artificial intelligence systems, now employed across a wide array of fields, can also serve as valuable aids for legal practitioners. Increasingly sophisticated tools enhance information search capabilities, automate the drafting or verification of legal documents, and facilitate technical evaluations, such as predictive justice. Utilizing such tools can yield significant benefits by enhancing the efficiency and quality of legal processes. In civil and common law systems, accessing legal judgments to retrieve legal decisions is essential for various legal tasks, including defending clients, constructing cases for prosecution, and issuing judicial decisions. In Italy, to ensure widespread information on the courts' decisions, for this purpose, a dedicated body, the Ufficio del Massimario, was established, whose purpose is to produce massime. In a concise yet detailed manner, these summaries (massime) encapsulate the legal principles articulated in judgments. Hence, legal professionals can consult these massime instead of delving into the entirety of legal decisions. The task of summarising legal texts and producing massime has been widely addressed in the last years <ref type="bibr" target="#b0">[1]</ref>, especially with the advent of the Generative AI <ref type="bibr" target="#b1">[2,</ref><ref type="bibr" target="#b2">3]</ref>. Given the complexity of the task, the approach outlined in <ref type="bibr" target="#b0">[1]</ref> focuses on handling the automatic production of a massima as an extractive summarization task. This involves extracting the most pertinent part of the judgement to assist in the drafting of the massima by the des-ignated office, utilizing a human-in-the-loop approach as discussed in <ref type="bibr" target="#b3">[4]</ref>. The process of analyzing judgments and extracting relevant sentences can be significantly simplified through the use of pre-trained models <ref type="bibr" target="#b4">[5,</ref><ref type="bibr" target="#b5">6]</ref>. These models function as versatile universal sentence/text encoders, capable of addressing a range of downstream tasks, including summarization <ref type="bibr" target="#b6">[7]</ref>. These models consistently outperform other approaches, particularly after fine-tuning or domain-adaptation <ref type="bibr" target="#b7">[8]</ref>. Despite the success of pre-trained transformers and LLMs in other summarization tasks <ref type="bibr" target="#b8">[9]</ref>, certain phenomena, such as hallucination in the generation of the text <ref type="bibr" target="#b9">[10]</ref>, the task of producing massime is still challenging for current extractive and abstractive summarization systems. Additionally, legal texts are often extensive, further increasing the summarization task's complexity. Identifying the portions of the text that contain the relevant information to be reported in the massime becomes challenging due to their length <ref type="bibr" target="#b10">[11]</ref>. In this paper, we present an approach to producing an extractive summary by exploiting the ability of an LLM to generate abstract summaries from a document. Our approach selects, from the abstract, the sentences that best match the sentences in the source document. This approach, described in Sec. 2, reduces the hallucination phenomena, achieving results in a zero-shot setting, described in Sec. 3 comparable with a model trained with a domain dataset.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Extractive Projection</head><p>It is well known that generative models, particularly when used in summarization systems, are prone to hallucination phenomena (see <ref type="bibr" target="#b11">[12]</ref> and references therein). In this case, new terms or, in worst scenarios, even informa- tion and facts not present in the original document are generated in the output summary. Several attempts have been made to try to tame such unwanted behaviour (for instance, see <ref type="bibr" target="#b12">[13]</ref>), which may lead to serious problems in sensitive domains. Given its specific lexicon, the vast amount of fixed forms and judicial references, the legal domain is very delicate and unsuitable for a straightforward application of generative systems. To overcome such a problem, we introduce what we refer to as extractive projection, meaning, a transformation mapping a generated text into sentences of the original document.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Model</head><p>Defining the source documents 𝑑 ∈ 𝐷, the abstractive summary as 𝑎 ∈ 𝐴 with 𝐷, 𝐴 respectively the space of documents and abstractive summaries, the summarization prompt 𝑝 ∈ 𝑃 . The generative summarization transformation is defined as:</p><formula xml:id="formula_0">𝐺 : 𝐷 × 𝑃 → 𝐴 𝑎 = 𝐺(𝑑|𝑝) .<label>(1)</label></formula><p>We introduce the extractive summary as 𝑎 ′ ∈ 𝐴 ′ ⊂ 𝐴, and the extractive projection</p><formula xml:id="formula_1">Γ Γ : 𝐷 ˜× 𝐴 → 𝐴 ′ 𝑎 ′ = Γ(𝑑 ˜; 𝑎) ,<label>(2)</label></formula><p>where 𝑑 ˜∈ 𝐷 ˜, and 𝐷 ˜is the space of segmented documents (i.e., containing the same documents as 𝐷, but each one is split into a set of segments).</p><p>The projection Γ used in this work is a slightly modified version of the algorithm proposed in <ref type="bibr" target="#b6">[7]</ref> to pre-process the data. As a main difference from <ref type="bibr" target="#b6">[7]</ref> we allow the algorithm to select up to all the segments present in the document, without any parameter fixed a priori. Moreover, while in <ref type="bibr" target="#b6">[7]</ref> this greedy selection algorithm is used to obtain an oracle summary for each document used as a reference to train the extractive model, here this algorithm is used to project the (abstractive) generated summary into the segments of the original document. Note that such procedure completely removes by construction any possibility of hallucination since the projection cuts off all possible novelties and generations. The greedy selection procedure employed is then simply a combinatorial optimization algorithm based on coverage metrics. In this respect, we tested several metrics, ranging from the average of ROUGE-1 and ROUGE-2 <ref type="bibr" target="#b13">[14]</ref> as originally proposed in <ref type="bibr" target="#b6">[7]</ref> to different linear combinations of Rouge-n and more sophisticated similarity metrics (e.g., BERTscore <ref type="bibr" target="#b14">[15]</ref>. We observe that with the exception of very rare cases where the generated summary is produced in a different language with respect to the original document, all the coverage metrics produce accurate results (see Tab. 1).</p><p>In the multilingual setup, only a similarity metric based on multilingual embeddings, which is insensitive to language shifts, produces reasonable results, while ROUGE does not work correctly.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Results</head><p>As discussed in Sec. 1, we trained and tested the extractive summarization systems introduced in [1] on a dataset composed by judgments and massime from different courts <ref type="foot" target="#foot_0">1</ref> . Starting from a whole dataset of 1340 couple of (judgement, massima), we randomly selected a subset of 199 of them as a validation set, 940 as a train set, and the remnant 201 as test set. The latter has been further refined down to 61 "high quality" examples. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 2</head><p>Prompts used for generic summarization (𝑝 1 ) and domain tuned task summarization (𝑝 2 ).</p><p>selection, we first used the greedy algorithm proposed in Sec. 2 based on the average of ROUGE1 and ROUGE2 and then selected only data with that value larger or equal to 0.6 (see Fig. <ref type="figure" target="#fig_3">2</ref>). The scores for the extractive model Ext, compared with the Oracle and those produced using generative models, are collected in Tab. 1. More specifically, we used two different prompts 𝑝1 and 𝑝2 (for details see Tab. 3) to estimate the effect of a "generic" summarization prompt, with a "task tuned" prompt specifically referring to the features of a massima <ref type="bibr" target="#b15">[16]</ref>. As expected, we observe a small improvement in scores with all generative models using 𝑝2 over 𝑝1. Moreover, we compare the scores of a straightforward abstractive summarization Abs, with the setup proposed in this work, i.e., including the extractive projection called Gen-Ext in Tab. 1. For all the evaluations, we used a generative model of the gpt-turbo <ref type="bibr" target="#b16">[17]</ref> family 2 . Interestingly, the scores obtained using zero-shots (no fine-tuning or contextual examples are involved) generative models, in both their types: abstractive (Abs) and extractive (Gen-Ext), seem to perform reasonably well when compared to the Ext model. An example of the summaries produced in all the setups are displayed in Tab. 3. It is worth noting that the scores obtained in this work should be interpreted only as a reference. They are affected by large statistical fluctuations, which make a direct comparison among the scores very tricky. Moreover, coverage scores are known to have a limited correlation with the effective quality of the summary produced, which requires some human evaluation by domain experts.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Conclusions</head><p>In this work we discussed the first results of a novel approach that can be used to obtain "hallucination"-free results out of a generative model. We applied such pro-  cedure in a legal domain, where preserving factuality is mandatory.</p><p>While obtain only partial results, we find them to be reasonably promising but requiring some further investigation.</p><p>A comparison with different "open source" LLMs as generative models, estimating the parameter scaling effects on the performances and a complete or partial (see for instance <ref type="bibr" target="#b17">[18]</ref> ) fine-tuning or domain-adaptation is dedicated to future studies. In conclusion, while discounting the difficulty of the task, given both by the inherent complexity of the structure of a massima that cannot be treated as a simple summary, and by the difficult evaluation of the results found, as well as the fact that automatic "token"-coverage metrics require some evaluation by human domain experts, we believe that LLMs, if appropriately applied, can offer a valuable tool even in domains where factuality is paramount.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Summary</head><p>Text Target L'inserzione automatica di clausole, prevista dall'art. 1339 cod. civ., costituisce una restrizione significativa del diritto di libertà economica consacrato dall'art. 41 Cost. di cui è espressione l'autonomia privata, e deve quindi trovare il suo fondamento in una legge formale o in un altro atto avente valore di legge in senso sostanziale o da esso richiamato tramite rinvio integrativo; pertanto, con riferimento al contratto avente ad oggetto la prestazione di servizi di vigilanza da parte di un istituto privato, tale integrazione non può aver luogo in base al decreto con cui il prefetto approva la relativa tariffa ai sensi dell'art. 135 del r.d. n. 773 del 1931, trattandosi di un mero atto amministrativo espressivo di un sindacato di congruità, avente natura di merito, del prezzo di una prestazione contrattuale, né l'ingerenza con efficacia reale sulla tariffa proposta dall'Istituto e liberamente accettata dal committente può trovare fondamento nella generale potestà del prefetto di impartire prescrizioni nell'interesse pubblico, in sede di rilascio di autorizzazioni di polizia, ai sensi dell'art. 9 del r.d. n. 773 cit.. L'autonomia privata non può essere derogata da atti amministrativi, ma solo da leggi formali. Il principio dell'autonomia privata è fondamentale e non può essere interpretato estensivamente. Le tariffe prefettizie non hanno efficacia normativa e non possono essere conferite da una presunta convergenza di vedute delle parti. Inoltre, il dovere di solidarietà sociale non può essere utilizzato come motivo di condanna al pagamento del maggior prezzo. La libertà economica può subire restrizioni solo in forza di una disposizione di legge. L'inserzione automatica di clausole ex art. 1339 c.c. riguarda solo i prezzi imposti con legge e non può essere derogata da atti amministrativi. La fedeltà alla decisione giuridica richiede il rispetto di tali principi.</p><p>Gen-Ext(p 1 ) -con atto di citazione notificato il 5 luglio 1995 l'istituto cooperativo di v. ed i. c. s.r.l. conveniva dinanzi al tribunale di lamezia terme l'istituto nazionale previdenza sociale per ottenerne la condanna al pagamento della somma di l. 639.590.300 a titolo di maggior corrispettivo per il servizio di vigilanza espletato in favore dell'ente, presso la sede di lamezia, dovuto in forza dell'inserzione automatica della tariffa inderogabile emanata dal prefetto di catanzaro in materia di vigilanza, superiore al prezzo pattuito con il contratto stipulato il 25 agosto 1983. dopo l'espletamento di consulenza tecnica d'ufficio il tribunale di lamezia terme con sentenza 23-27 novembre 1999 condannava l'inps al pagamento della somma di l. 1.148.787.862, oltre le spese di giudizio. avverso la sentenza, non notificata, proponeva ricorso per cassazione il fallimento dell'istituto cooperativo di v. e i. 157 del relativo regolamento di esecuzione, nonché l'art. 1175 e 1375 cod. civ. e dell'art. diritto -con il primo motivo il ricorrente deduce la violazione degli artt. 9, 134 e 135 del citato testo unico delle leggi di pubblica sicurezza e dell'art. 1339 cod. civile. con il secondo motivo ricorrente censura l'omessa motivazione nel discostarsi dalla concorde interpretazione delle parti. con l'ultimo motivo il fallimento deduce la violazione degli artt. 2 della costituzione.</p><p>Gen-Ext(p 2 ) 1339, cod. civ. 1339 cod. civ. l'inserzione automatica di clausole, prevista dall'art. 1339, cod. civ., costituisce una deroga incisiva al principio dell'autonomia privata e deve quindi trovare il suo fondamento in una legge formale -come testualmente previsto dalla norma -o in altro atto avente valore di legge in senso sostanziale o da esso richiamato tramite rinvio integrativo. il diritto di libertà economica consacrato dall'art. 41 cost., di cui è espressione l'autonomia negoziale delle parti nel modellare il contenuto di un contratto, può soffrire restrizioni solo in forza di una disposizione di legge, insuscettibile di interpretazioni estensive (ibidem, terzo comma). ne consegue la vigenza, in subiecta materia, di un principio di stretta interpretazione dell'art. civ. ; vieppiù giustificato da esigenze di tutela della concorrenza e del mercato, che verrebbero lese da una pratica di prezzi amministrati. l'asserita convergenza di vedute sull'efficacia cogente delle tariffe prefettizie non può, neanche in astratto, valere a conferire loro l'efficacia normativa di cui sono intrinsecamente prive. l'invocazione di un inderogabile dovere di solidarietà sociale che avrebbe imposto la maggiorazione del prezzo non ha, infatti, alcuna attinenza con l'operatività dell'eterointegrazione ex art.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 3</head><p>Example summary comparison among the various summarization systems proposed.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Sketch of the extractive summarization system proposed in this work.</figDesc><graphic coords="2,302.62,84.19,204.19,128.96" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head></head><label></label><figDesc>in Italian of 150 words of the following text delimited by triple backquote: "'content"'𝑝2Scrivi una massima in Italiano di 150 parole della seguente porzione di testo delimitata dalle virgolette. La massima deve rispondere ai seguenti generali requisiti: a) fedeltà alla decisione; b) sintesi nell'enunciazione del principio; c) chiarezza e precisione del principio enunciato La massima costituisce l'enucleazione del principio di diritto e non il riassunto della decisione e non può tradursi nella mera riproduzione di passaggi argomentativi della motivazione. "'content"'</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>2 gpt- 3 . 5 -</head><label>35</label><figDesc>turbo-1106.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: Fraction of test data as a function of the score (ROUGE 1 +ROUGE 2 )/2 computed on the segments extracted by the oracle combinatorial algorithm.</figDesc><graphic coords="3,302.62,255.73,204.17,155.88" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1 Mean</head><label>1</label><figDesc>ROUGE𝑛-𝑓 1 scores computed on test data for different models. Ext is an extractive model trained on the Oracle. Gen-Ext and Abs are the models based on pure abstractive summarization with and w/o extractive projection respectively. Results with different prompts 𝑝 1 (generic summarization prompt) and 𝑝 2 (domain tuned summarization prompt) are also displayed explicitly in Tab. 3</figDesc><table><row><cell></cell><cell>ROUGE 1</cell><cell>ROUGE 2</cell><cell>ROUGE 3</cell></row><row><cell>Oracle</cell><cell>0.81</cell><cell>0.71</cell><cell>0.65</cell></row><row><cell>Ext</cell><cell>0.40</cell><cell>0.30</cell><cell>0.28</cell></row><row><cell>Abs(𝑝 1 )</cell><cell>0.32</cell><cell>0.10</cell><cell>0.05</cell></row><row><cell>Gen-Ext(𝑝 1 )</cell><cell>0.31</cell><cell>0.12</cell><cell>0.08</cell></row><row><cell>Abs(𝑝 2 )</cell><cell>0.35</cell><cell>0.13</cell><cell>0.07</cell></row><row><cell>Gen-Ext(𝑝 2 )</cell><cell>0.38</cell><cell>0.20</cell><cell>0.16</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head></head><label></label><figDesc>Abs(p 1 ) L'Istituto cooperativo di V. ed I. C. s.r.l. ha citato in giudizio l'Istituto Nazionale Previdenza Sociale per ottenere il pagamento di una somma di denaro a titolo di maggior corrispettivo per il servizio di vigilanza. Dopo un lungo iter giudiziario, la Corte d'appello di Catanzaro ha respinto la domanda, sostenendo che il provvedimento prefettizio in materia di tariffe per la vigilanza non poteva derogare al contenuto contrattuale liberamente scelto dalle parti. Il fallimento dell'Istituto Cooperativo di V. e I. C. ha proposto ricorso per cassazione, ma la Corte ha respinto il ricorso, condannando al pagamento delle spese di giudizio. La Corte ha sostenuto che l'inserzione automatica di clausole ex art. 1339 c.c. riguardava solo i prezzi imposti con legge e che l'ingerenza con efficacia reale sulla tariffa proposta dall'istituto di vigilanza e liberamente accettata dal committente non poteva trovare fondamento nella generale potestà di impartire prescrizioni nell'interesse pubblico.</figDesc><table><row><cell>Abs(p 2 )</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">The data are publicly available on the website https://www.inps.it/it/it/inps-comunica/atti/sentenze.html</note>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Legal summarization: to each court its own model</title>
		<author>
			<persName><forename type="first">F</forename><surname>Achena</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Preti</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Venditti</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Ranaldi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Giannone</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">M</forename><surname>Zanzotto</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Favalli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Romagnoli</surname></persName>
		</author>
		<ptr target="https://ceur-ws.org/Vol-3596/paper1.pdf" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 9th Italian Conference on Computational Linguistics</title>
		<title level="s">CEUR Workshop Proceedings</title>
		<editor>
			<persName><forename type="first">F</forename><surname>Boschetti</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><forename type="middle">E</forename><surname>Lebani</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">B</forename><surname>Magnini</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Novielli</surname></persName>
		</editor>
		<meeting>the 9th Italian Conference on Computational Linguistics<address><addrLine>Venice, Italy, Novem</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2023">ber 30 -December 2, 2023. 2023</date>
			<biblScope unit="volume">3596</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<monogr>
		<author>
			<persName><forename type="first">T</forename><surname>Pont</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Galli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Loreggia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Pisano</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Rovatti</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Sartor</surname></persName>
		</author>
		<idno type="DOI">10.48550/arXiv.2308.04416</idno>
		<idno type="arXiv">arXiv:2308.04416</idno>
		<idno>arXiv:2308.04416</idno>
		<title level="m">Legal Summarisation through LLMs: The PRODIGIT Project</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Summarization di testi giuridici: una sperimentazione con gpt-3</title>
		<author>
			<persName><forename type="first">M</forename><surname>Cherubini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Romano</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Bolioli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>De Francesco</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Benedetto</surname></persName>
		</author>
		<idno type="DOI">10.32091/RIID0103</idno>
	</analytic>
	<monogr>
		<title level="j">Rivista Italiana di Informatica e Diritto</title>
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Viewpoint: Human-in-the-loop artificial intelligence</title>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">M</forename><surname>Zanzotto</surname></persName>
		</author>
		<idno type="DOI">10.1613/jair.1.11345</idno>
		<ptr target="https://doi.org/10.1613%2Fjair.1.11345.doi:10.1613/jair.1.11345" />
	</analytic>
	<monogr>
		<title level="j">Journal of Artificial Intelligence Research</title>
		<imprint>
			<biblScope unit="volume">64</biblScope>
			<biblScope unit="page" from="243" to="252" />
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<monogr>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">E</forename><surname>Peters</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Neumann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Iyyer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Gardner</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Clark</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Zettlemoyer</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1802.05365</idno>
		<title level="m">Deep contextualized word representations</title>
				<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">BERT: Pre-training of deep bidirectional transformers for language understanding</title>
		<author>
			<persName><forename type="first">J</forename><surname>Devlin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M.-W</forename><surname>Chang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Toutanova</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/N19-1423</idno>
		<ptr target="https://aclanthology.org/N19-1423.doi:10.18653/v1/N19-1423" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</title>
		<title level="s">Long and Short Papers</title>
		<meeting>the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies<address><addrLine>Minneapolis, Minnesota</addrLine></address></meeting>
		<imprint>
			<publisher>Association for Computational Linguistics</publisher>
			<date type="published" when="2019">2019</date>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page" from="4171" to="4186" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<monogr>
		<title level="m" type="main">Text summarization with pretrained encoders</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Lapata</surname></persName>
		</author>
		<idno type="DOI">10.48550/ARXIV.1908.08345</idno>
		<ptr target="https://arxiv.org/abs/1908.08345.doi:10.48550/ARXIV.1908.08345" />
		<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Lifelong pretraining: Continually adapting language models to emerging corpora</title>
		<author>
			<persName><forename type="first">X</forename><surname>Jin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Zhu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Xiao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S.-W</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Wei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Arnold</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Ren</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/2022.bigscience-1.1</idno>
		<ptr target="https://aclanthology.org/2022.bigscience-1.1.doi:10.18653/v1/2022.bigscience-1.1" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of BigScience Episode #5 -Workshop on Challenges &amp; Perspectives in Creating Large Language Models</title>
		<title level="s">Association for Computational Linguistics</title>
		<meeting>BigScience Episode #5 -Workshop on Challenges &amp; Perspectives in Creating Large Language Models<address><addrLine>virtual+Dublin</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2022">2022</date>
			<biblScope unit="page" from="1" to="16" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Benchmarking Large Language Models for News Summarization</title>
		<author>
			<persName><forename type="first">T</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Ladhak</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Durmus</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Liang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Mckeown</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">B</forename><surname>Hashimoto</surname></persName>
		</author>
		<idno type="DOI">10.1162/tacl_a_00632</idno>
		<ptr target="https://doi.org/10.1162/tacl_a_00632.doi:10.1162/tacl_a_00632" />
	</analytic>
	<monogr>
		<title level="j">Transactions of the Association for Computational Linguistics</title>
		<imprint>
			<biblScope unit="volume">12</biblScope>
			<biblScope unit="page" from="39" to="57" />
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Improving abstractive summarization of legal rulings through textual entailment</title>
		<author>
			<persName><forename type="first">D</forename><surname>Vargas Feijó</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><forename type="middle">P</forename><surname>Moreira</surname></persName>
		</author>
		<idno type="DOI">10.1007/S10506-021-09305-4</idno>
		<ptr target="https://doi.org/10.1007/s10506-021-09305-4.doi:10.1007/S10506-021-09305-4" />
	</analytic>
	<monogr>
		<title level="j">Artif. Intell. Law</title>
		<imprint>
			<biblScope unit="volume">31</biblScope>
			<biblScope unit="page" from="91" to="113" />
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<monogr>
		<author>
			<persName><forename type="first">E</forename><surname>Bauer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Stammbach</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Gu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Ash</surname></persName>
		</author>
		<title level="m">Legal extractive summarization of u</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
	<note>court opinions</note>
</biblStruct>

<biblStruct xml:id="b11">
	<monogr>
		<author>
			<persName><forename type="first">L</forename><surname>Huang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Yu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Ma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Zhong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Feng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Peng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Feng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Qin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Liu</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2311.05232</idno>
		<title level="m">A survey on hallucination in large language models: Principles, taxonomy, challenges, and open questions</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<monogr>
		<title level="m" type="main">Towards mitigating hallucination in large language models via self-reflection</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Ji</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Yu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Ishii</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Fung</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2310.06271</idno>
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">ROUGE: A package for automatic evaluation of summaries</title>
		<author>
			<persName><forename type="first">C.-Y</forename><surname>Lin</surname></persName>
		</author>
		<ptr target="https://aclanthology.org/W04-1013" />
	</analytic>
	<monogr>
		<title level="m">Text Summarization Branches Out, Association for Computational Linguistics</title>
				<meeting><address><addrLine>Barcelona, Spain</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2004">2004</date>
			<biblScope unit="page" from="74" to="81" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<monogr>
		<author>
			<persName><forename type="first">T</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Kishore</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><forename type="middle">Q</forename><surname>Weinberger</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Artzi</surname></persName>
		</author>
		<idno>CoRR abs/1904.09675</idno>
		<ptr target="http://arxiv.org/abs/1904.09675.arXiv:1904.09675" />
		<title level="m">Bertscore: Evaluating text generation with BERT</title>
				<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<monogr>
		<author>
			<persName><forename type="first">C</forename><surname>Di Cassazione</surname></persName>
		</author>
		<ptr target="https://www.cortedicassazione.it/resources/cms/documents/SINTESI_CRITERI_DELLA_MASSIMAZIONE_CIVILE_E_PENALE.pdf" />
		<title level="m">Sintesi criteri della massimazione civile e penale</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<monogr>
		<title/>
		<author>
			<persName><surname>Openai</surname></persName>
		</author>
		<idno>Gpt-3.5-turbo-1106</idno>
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
	<note>large language model</note>
</biblStruct>

<biblStruct xml:id="b17">
	<monogr>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">J</forename><surname>Hu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Shen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Wallis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Allen-Zhu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Chen</surname></persName>
		</author>
		<idno>CoRR abs/2106.09685</idno>
		<ptr target="https://arxiv.org/abs/2106.09685.arXiv:2106.09685" />
		<title level="m">Lora: Low-rank adaptation of large language models</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
